Memorization vs. Reasoning: Updating LLMs with New Knowledge

Aochong Oliver Li, Tanya Goyal


Abstract
Large language models (LLMs) encode vast amounts of pre-trained knowledge in their parameters, but updating them as real-world information evolves remains a challenge. Existing methodologies and benchmarks primarily target entity substitutions, failing to capture the full breadth of complex real-world dynamics. In this paper, we introduce Knowledge Update Playground (KUP), an automatic pipeline for simulating realistic knowledge updates reflected in an evidence corpus. KUP’s evaluation framework includes direct and indirect probes to both test memorization of updated facts and reasoning over them, for any update learning methods. Next, we present a lightweight method called memory conditioned training (MCT), which conditions tokens in the update corpus on self-generated ”memory” tokens during training. Our strategy encourages LLMs to surface and reason over newly memorized knowledge at inference. Our results on two LLM families show that (1) KUP benchmark is highly challenging, with the best CPT models achieving <2% in indirect probing setting (reasoning) and (2) MCT training significantly outperforms prior continued pre-training (CPT) baselines, improving direct probing (memorization) results by up to 25.4%.
Anthology ID:
2025.findings-acl.1326
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venues:
Findings | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
25853–25874
Language:
URL:
https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.1326/
DOI:
Bibkey:
Cite (ACL):
Aochong Oliver Li and Tanya Goyal. 2025. Memorization vs. Reasoning: Updating LLMs with New Knowledge. In Findings of the Association for Computational Linguistics: ACL 2025, pages 25853–25874, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Memorization vs. Reasoning: Updating LLMs with New Knowledge (Li & Goyal, Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.1326.pdf