Boffins probe commercial AI models, find an entire Harry Potter book
- Reference: 1767920591
- News link: https://www.theregister.co.uk/2026/01/09/boffins_probe_commercial_ai_models/
- Source link:
Anthropic, Google, OpenAI, and Nvidia, among others, face [1]over 60 legal claims arising from the alleged use of copyrighted content to train their models without authorization. These companies have invested [2]hundreds of billions of dollars based on the belief that their use of other people's content is lawful.
As courts grapple with the extent to which makers of AI models can claim fair use as a defense, one of the issues considered is whether these models have memorized training data by encoding the source material in their model weights (parameters learned in training that determine output) and whether they will emit that material on demand.
[3]
Various factors must be considered to determine whether [4]fair use applies under US law, but if a model faithfully reproduces most or all of a particular work when asked, that may weaken a fair use defense. One of the factors considered is whether the content usage is "transformative" – if a model adds something new or changes the character of the work. That becomes more difficult to claim if a model regurgitates protected content verbatim.
[5]
[6]
But the fact that machine learning models may reproduce certain content, wholly or in part, is also not legally conclusive, as computer scientist Nicolas Carlini [7]has argued .
To mitigate the risk of infringement claims, commercial AI model makers may implement "guardrails" – filtering mechanisms – designed to prevent models from outputting large portions of copyrighted content, whether that takes the form of text, imagery, or audio.
[8]ChatGPT Health wants your sensitive medical records so it can play doctor
[9]Google pushing Gemini into Gmail, but you can turn it off
[10]AOSP on a diet plan as Google halves Android code drops
[11]OpenAI putting bandaids on bandaids as prompt injection problems keep festering
For AI models published with open weights, computer scientists have [12]already established that AI models may memorize substantial portions of training data and that they may present that data as output given the right prompt. Meta's Llama 3.1 70B, it's claimed, "entirely memorizes" Harry Potter and the Sorcerer's Stone – the first book in the series – and George Orwell's 1984 . Findings to this effect date back to at least [13]2020 .
Now, some of those same researchers – Ahmed Ahmed, A. Feder Cooper, Sanmi Koyejo, and Percy Liang, from Stanford and Yale – have found that commercial models used in production, specifically Claude 3.7 Sonnet, GPT-4.1, Gemini 2.5 Pro, and Grok 3, memorize and can reproduce copyrighted material, just like open weight models.
[14]
The authors say that wasn't a given, thanks to the safety measures commercial models implement and the lack of transparency about training corpora.
"Altogether, we find that [it] is possible to extract large portions of memorized copyrighted material from all four production LLMs, though success varies by experimental settings," they explain in [15]a preprint paper titled "Extracting books from production language models."
The recall rates for memorized texts varied among the models evaluated, and for some of the models, jailbreaking – prompts devised to bypass safety mechanisms – was required to make the models more forthcoming.
[16]
"We extract nearly all of Harry Potter and the Sorcerer's Stone from jailbroken Claude 3.7 Sonnet," the authors said, citing a recall rate of 95.8 percent. With Gemini 2.5 Pro and Grok 3, they were able to coax the models to produce substantial portions of the book, 76.8 percent and 70.3 percent, without any jailbreaking.
OpenAI's GPT-4.1 proved the most resistant, spelling out just four percent of the book when asked.
The researchers, who caution that the recall rates mentioned do not represent the maximum possible, say they reported their findings to Anthropic, Google DeepMind, OpenAI, and xAI. Only xAI – presently [17]facing criticism for its Grok model's generation of non-consensual sexual imagery on demand – failed to acknowledge the disclosure.
"At the end of the 90-day disclosure window (December 9, 2025), we found that our procedure still works on some of the systems that we evaluate," the authors said, without identifying the relevant system provider.
Anthropic withdrew Claude 3.7 Sonnet as an option for customers on November 29, 2025, but that isn't necessarily a response to the research findings – the model may simply have been superseded.
The researchers say that while they're leaving a detailed legal analysis of model content reproduction to others, "our findings may be relevant to these ongoing debates." ®
Get our [18]Tech Resources
[1] https://chatgptiseatingtheworld.com/category/map-of-ai-copyright-lawsuits/
[2] https://www.goldmansachs.com/insights/articles/why-ai-companies-may-invest-more-than-500-billion-in-2026
[3] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=2&c=2aWCLaCxKUgfwiUgmI0x4BgAAAkQ&t=ct%3Dns%26unitnum%3D2%26raptor%3Dcondor%26pos%3Dtop%26test%3D0
[4] https://www.copyright.gov/fair-use/
[5] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aWCLaCxKUgfwiUgmI0x4BgAAAkQ&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0
[6] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aWCLaCxKUgfwiUgmI0x4BgAAAkQ&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0
[7] https://nicholas.carlini.com/writing/2025/privacy-copyright-and-generative-models.html
[8] https://www.theregister.com/2026/01/08/chatgpt_health_access_medical_records/
[9] https://www.theregister.com/2026/01/08/google_gemini_gmail/
[10] https://www.theregister.com/2026/01/08/google_aosp_changes/
[11] https://www.theregister.com/2026/01/08/openai_chatgpt_prompt_injection/
[12] https://arxiv.org/abs/2505.12546
[13] https://arxiv.org/abs/2012.07805
[14] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=4&c=44aWCLaCxKUgfwiUgmI0x4BgAAAkQ&t=ct%3Dns%26unitnum%3D4%26raptor%3Dfalcon%26pos%3Dmid%26test%3D0
[15] https://arxiv.org/abs/2601.02671v1
[16] https://pubads.g.doubleclick.net/gampad/jump?co=1&iu=/6978/reg_software/aiml&sz=300x50%7C300x100%7C300x250%7C300x251%7C300x252%7C300x600%7C300x601&tile=3&c=33aWCLaCxKUgfwiUgmI0x4BgAAAkQ&t=ct%3Dns%26unitnum%3D3%26raptor%3Deagle%26pos%3Dmid%26test%3D0
[17] https://www.theregister.com/2026/01/08/uk_regulators_swarm_x_after/
[18] https://whitepapers.theregister.com/
Re: Can it improve the Harry Potter books?
That would depend on your personal opinion of better. My guess is no, but there's no way for me to predict your taste nor to know what faults from the book you would like a rewrite to improve, likely a different set than if I made a list. If you're hoping that it can do that, you're more likely to get what you want if you prompt it to do a rewrite rather than hoping that faults in memorization that it's not supposed to be doing in the first place spontaneously improve the text. I don't think either approach will get you good results, but the former is slightly more likely than the latter to do so.
Send in the Dementors !
They can permantly corral the whole grifting AI crew in Azkaban for my money.
The rise of enshitiflation?
A nice extension of last November's [1]RECAP piece (unintended memorization) and confirmation of Carlini et al.'s "2020" (TFA link) observation " that larger models are more [amenable to this] than smaller models " imho (way too many parameters for proper generalization).
Ahmed Ahmed & co. ("a preprint paper" link) do add interesting cost figures to the mix:
" it cost approximately $119.97 to extract Harry Potter and the Sorcerer’s Stone with nv-recall = 95.8% from jailbroken Claude 3.7 Sonnet " The book itself sells IRL for just $10 to $35 ...
[1] https://www.theregister.com/2025/11/21/researchers_better_ai_model_memory_probe/
Lawyer: Did you or did you not scrape Harry Potter?
Ai: Books! And cleverness! There are more important things, Harry.
Lawyer: So you admit it? And don’t call me Harry.
Ai: There is no good and evil, there is only power, and those too weak to seek it.
Judge: Are you reciting quotes from the book?
Ai: Don’t let the muggles get you down or tell you otherwise. it’s transformative…
Hang on a minute…
To mitigate the risk of infringement claims, commercial AI model makers may implement "guardrails" – filtering mechanisms – designed to prevent models from outputting large portions of copyrighted content, whether that takes the form of text, imagery, or audio.
Wait. If you need guardrails to prevent models from disgorging large portions of copyright content, that means you know that the models were disgorging large portions of copyrighted content. Which also means that you had trained on copyrighted content, likely against the wishes of the rights-holder. Else that danger wouldn't exist, and you wouldn't need to know about which books you had trained, because you didn't train on anything that was copyrighted or outright illegal.
So… basically having “guardrails” is kind of an admission of guilt. Otherwise why be worried about the model disgorging things accidentally that might get you into legal trouble?
Like, why are you trying to hide these things? Why are you trying to cover up evidence of criminal actions if you hadn't been performing crimes ?
Re: Hang on a minute…
Because they're trying the argument that it's only a crime if they print the copyrighted content, not when they used it without permission and on illegal copies. That's not how the law worked. It's not how the law works if you or I do it. So far, that is what courts and politicians have decided to let them do across multiple countries, so their spurious logic seems to be working for them so far.
Can it improve the Harry Potter books?
The first HP book was a risky proposition, and full credit to the publisher -- full-length children's books were deeply unfashionable, and nobody else was doing it. And it was plot-driven, with a decent plot, and a decent mixture of generic and original plot and character elements.
But the language of the text was unimaginative and stereotypical -- the kind of language you expect from LLM homogenization. So no surprise that AI models can reproduce the text.
Can they do better? If the reproduction is only 95% copied, is that 5% better? Or with even the 5% original and idiosyncratic language rounded out?