Citation Drift: Measuring Reference Stability in Multi-Turn LLM Conversations

Gokul Srinath Seetha Ram


Abstract
Large Language Models (LLMs) are increasingly used for scientific writing and research assistance, yet their ability to maintain consistent citations across multi-turn conversations remains unexplored. This paper introduces the concept of citation drift—the phenomenon where references mutate, disappear, or get fabricated during extended LLM interactions. We analyze 240 conversations across four LLaMA models using 36 authentic scientific papers from six domains and find significant citation instability. LLaMA-4-Maverick-17B achieves the highest stability (0.481) and lowest fabrication entropy, while LLaMA-4-Scout-17B fabricates up to 85.6% of citations. We introduce five new metrics—stability, fabrication rate, drift rate, drift entropy, and willingness-to-cite—providing a standardized framework for evaluating factual reliability in scientific dialogue systems. Our benchmark offers reproducible, model-agnostic evaluation tools for assessing citation reliability in AI-assisted research workflows.
Anthology ID:
2025.wasp-main.20
Volume:
Proceedings of the Third Workshop for Artificial Intelligence for Scientific Publications
Month:
December
Year:
2025
Address:
Mumbai, India and virtual
Editors:
Alberto Accomazzi, Tirthankar Ghosal, Felix Grezes, Kelly Lockhart
Venues:
WASP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
186–191
Language:
URL:
https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.wasp-main.20/
DOI:
Bibkey:
Cite (ACL):
Gokul Srinath Seetha Ram. 2025. Citation Drift: Measuring Reference Stability in Multi-Turn LLM Conversations. In Proceedings of the Third Workshop for Artificial Intelligence for Scientific Publications, pages 186–191, Mumbai, India and virtual. Association for Computational Linguistics.
Cite (Informal):
Citation Drift: Measuring Reference Stability in Multi-Turn LLM Conversations (Ram, WASP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.wasp-main.20.pdf