@inproceedings{baeumel-etal-2025-lookahead,
title = "The Lookahead Limitation: Why Multi-Operand Addition is Hard for {LLM}s",
author = "Baeumel, Tanja and
Genabith, Josef Van and
Ostermann, Simon",
editor = "Belinkov, Yonatan and
Mueller, Aaron and
Kim, Najoung and
Mohebbi, Hosein and
Chen, Hanjie and
Arad, Dana and
Sarti, Gabriele",
booktitle = "Proceedings of the 8th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.blackboxnlp-1.15/",
pages = "250--262",
ISBN = "979-8-89176-346-3",
abstract = "Autoregressive large language models (LLMs) exhibit impressive performance across various tasks but struggle with simple arithmetic, such as additions of two or more operands. We show that this struggle arises from LLMs' use of a simple one-digit lookahead heuristic, which forms an upper bound for LLM performance and accounts for characteristic error patterns in two-operand addition and failure in multi-operand addition, where the carry-over logic is more complex. Our probing experiments and digit-wise accuracy evaluation show that the evaluated LLMs fail precisely where a one-digit lookahead is insufficient to account for cascading carries. We analyze the impact of tokenization strategies on arithmetic performance and show that all investigated models, regardless of tokenization and size, are inherently limited in the addition of multiple operands due to their reliance on a one-digit lookahead heuristic. Our findings reveal limitations that prevent LLMs from generalizing to more complex numerical reasoning."
}Markdown (Informal)
[The Lookahead Limitation: Why Multi-Operand Addition is Hard for LLMs](https://preview.aclanthology.org/ingest-emnlp/2025.blackboxnlp-1.15/) (Baeumel et al., BlackboxNLP 2025)
ACL