Memorization: A Close Look at Books

Iris Ma, Ian Domingo, Alberto Krone-Martins, Pierre Baldi, Cristina Lopes


Abstract
To what extent can entire books be extracted from LLMs? Using the Llama 3 70B family of models, and the “prefix-prompting” extractiontechnique, we were able to auto-regressively reconstruct, with a very high level of similarity, one entire book (Alice’s Adventures in Wonderland) from just the first 500 tokens. We were also able to obtain high extraction rates on several other books, piece-wise. However, these successes do not extend uniformly to all books. We show that extraction rates of books correlate with book popularity and thus, likely duplication in the training data. We also confirm the undoing of mitigations in the instruction-tuned Llama 3.1, following recent work (Nasr et al., 2025). We further find that this undoing comes from changes to only a tiny fraction of weights concentrated primarily in the lower transformer blocks. Our results provide evidence of the limits of current regurgitation mitigation strategies and introduce a framework for studying how fine-tuning affects the retrieval of verbatim memorization in aligned LLMs.
Anthology ID:
2025.l2m2-1.13
Volume:
Proceedings of the First Workshop on Large Language Model Memorization (L2M2)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Robin Jia, Eric Wallace, Yangsibo Huang, Tiago Pimentel, Pratyush Maini, Verna Dankers, Johnny Wei, Pietro Lesci
Venues:
L2M2 | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
169–182
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.l2m2-1.13/
DOI:
Bibkey:
Cite (ACL):
Iris Ma, Ian Domingo, Alberto Krone-Martins, Pierre Baldi, and Cristina Lopes. 2025. Memorization: A Close Look at Books. In Proceedings of the First Workshop on Large Language Model Memorization (L2M2), pages 169–182, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Memorization: A Close Look at Books (Ma et al., L2M2 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.l2m2-1.13.pdf