Language models can learn implicit multi-hop reasoning, but only if they have lots of training data

Yuekun Yao, Yupei Du, Dawei Zhu, Michael Hahn, Alexander Koller


Abstract
Implicit reasoning is the ability of a language model to solve multi-hop reasoning tasks in a single forward pass, without chain of thought.We investigate this capability using GPT2-style language models trained from scratch on controlled k-hop reasoning datasets (k = 2, 3, 4). We show that while such models can indeed learn implicit k-hop reasoning,the required training data grows exponentially in k, and the requirednumber of transformer layers grows linearly in k.We offer a theoretical explanation for why this depth growth is necessary.We further find that the data requirement can be mitigated, but not eliminated,through curriculum learning.
Anthology ID:
2025.emnlp-main.490
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
9695–9713
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.490/
DOI:
Bibkey:
Cite (ACL):
Yuekun Yao, Yupei Du, Dawei Zhu, Michael Hahn, and Alexander Koller. 2025. Language models can learn implicit multi-hop reasoning, but only if they have lots of training data. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 9695–9713, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Language models can learn implicit multi-hop reasoning, but only if they have lots of training data (Yao et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.490.pdf
Checklist:
 2025.emnlp-main.490.checklist.pdf