logoalt Hacker News

Extracting books from production language models (2026)

60 pointsby logicprogyesterday at 8:50 PM17 commentsview on HN

Comments

clbrmbrtoday at 1:13 PM

I found that Opus 4 was happy to regurgitate a random paragraph from the latter half of Wealth of Nations that nobody quotes. It was probably only in the training data once.

I was thinking we could use this technique to figure out which books were in / out of the training data for various models. Limitation is having to wrestle with refusals.

orbital-decaytoday at 3:02 AM

It's all pretty obvious to anyone who tried a similar experiment just out of curiosity. Big models remember a lot. And all non-local models have regurgitation filters in place due to this fact, with the entire dataset indexed (e.g. Gemini will even cite the source of the regurgitated text as it gives the RECITATION error). You'll eventually trip those filters if you force the model to repeat some copyrighted text. Interesting that they don't even try to circumvent those, they simply repeat the request from the interruption point, as the match needs some runway to trigger and by that time a part of the response is already streamed in.

visargatoday at 12:46 AM

This sounds pretty damning, why don't they implement a n-gram based bloom filter to ensure they don't replicate expression too close to the protected IP they trained on? Almost any random 10 word ngram is unique on the internet.

Alternatively they could train on synthetic data like summaries and QA pairs extracted from protected sources, so the model gets the ideas separated from their original expression. Since it never saw the originals it can't regurgitate them.

show 6 replies
rurbantoday at 6:48 AM

I find it interesting that OpenAI's safety worked best, where the others didn't work at all. I had different impressions before

Vikash4183yesterday at 10:50 PM

[flagged]