Evaluating LLMs for Quotation Attribution in Literary Texts: A Case Study of LLaMa3

Gaspard Michel, Elena V. Epure, Romain Hennequin, Christophe Cerisara


Abstract
Large Language Models (LLMs) have shown promising results in a variety of literary tasks, often using complex memorized details of narration and fictional characters. In this work, we evaluate the ability of Llama-3 at attributing utterances of direct-speech to their speaker in novels. The LLM shows impressive results on a corpus of 28 novels, surpassing published results with ChatGPT and encoder-based baselines by a large margin. We then validate these results by assessing the impact of book memorization and annotation contamination.We found that these types of memorization do not explain the large performance gain, making Llama-3 the new state-of-the-art for quotation attribution in English literature. We release publicly our code and data.
Anthology ID:
2025.naacl-short.62
Volume:
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers)
Month:
April
Year:
2025
Address:
Albuquerque, New Mexico
Editors:
Luis Chiruzzo, Alan Ritter, Lu Wang
Venue:
NAACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
742–755
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.naacl-short.62/
DOI:
Bibkey:
Cite (ACL):
Gaspard Michel, Elena V. Epure, Romain Hennequin, and Christophe Cerisara. 2025. Evaluating LLMs for Quotation Attribution in Literary Texts: A Case Study of LLaMa3. In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers), pages 742–755, Albuquerque, New Mexico. Association for Computational Linguistics.
Cite (Informal):
Evaluating LLMs for Quotation Attribution in Literary Texts: A Case Study of LLaMa3 (Michel et al., NAACL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.naacl-short.62.pdf