@inproceedings{geng-etal-2025-great,
title = "Great Memory, Shallow Reasoning: Limits of $k${NN}-{LM}s",
author = "Geng, Shangyi and
Zhao, Wenting and
Rush, Alexander M",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers)",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/Ingest-2025-COMPUTEL/2025.naacl-short.40/",
pages = "471--482",
ISBN = "979-8-89176-190-2",
abstract = "K-nearest neighbor language models (kNN-LMs), which integrate retrieval with next-word prediction, have demonstrated strong performance in language modeling as well as some downstream NLP benchmarks. These results have led researchers to argue that models trained on poor quality or outdated data could perform well by employing a kNN extension that has access to a higher-quality datastore. In this work, we ask whether this improved ability to recall information really translates into downstream abilities. We extensively evaluate kNN-LMs on a diverse set of tasks, ranging from sentiment classification and commonsense reasoning to multi-hop reasoning. Results show that kNN-LMs excel at memory-intensive tasks, where utilizing the patterns in the input is sufficient for determining the output, but struggle with reasoning tasks that require integrating multiple pieces of information to derive new knowledge. We further demonstrate through oracle experiments and qualitative analysis that even with perfect retrieval, kNN-LMs still fail to determine the correct answers, placing an upper bound on their reasoning performance."
}
Markdown (Informal)
[Great Memory, Shallow Reasoning: Limits of kNN-LMs](https://preview.aclanthology.org/Ingest-2025-COMPUTEL/2025.naacl-short.40/) (Geng et al., NAACL 2025)
ACL
- Shangyi Geng, Wenting Zhao, and Alexander M Rush. 2025. Great Memory, Shallow Reasoning: Limits of kNN-LMs. In Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 2: Short Papers), pages 471–482, Albuquerque, New Mexico. Association for Computational Linguistics.