What is an “Abstract Reasoner”? Revisiting Experiments and Arguments about Large Language Models

Tian Yun, Chen Sun, Ellie Pavlick


Abstract
Recent work has argued that large language models (LLMs) are not “abstract reasoners”, citing their poor zero-shot performance on a variety of challenging tasks as evidence. We revisit these experiments in order to add nuance to the claim. First, we show that while LLMs indeed perform poorly in a zero-shot setting, even tuning a small subset of parameters for input encoding can enable near-perfect performance. However, we also show that this finetuning does not necessarily transfer across datasets. We take this collection of empirical results as an invitation to (re-)open the discussion of what it means to be an “abstract reasoner”, and why it matters whether LLMs fit the bill.
Anthology ID:
2025.conll-1.11
Volume:
Proceedings of the 29th Conference on Computational Natural Language Learning
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Gemma Boleda, Michael Roth
Venues:
CoNLL | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
156–168
Language:
URL:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.conll-1.11/
DOI:
Bibkey:
Cite (ACL):
Tian Yun, Chen Sun, and Ellie Pavlick. 2025. What is an “Abstract Reasoner”? Revisiting Experiments and Arguments about Large Language Models. In Proceedings of the 29th Conference on Computational Natural Language Learning, pages 156–168, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
What is an “Abstract Reasoner”? Revisiting Experiments and Arguments about Large Language Models (Yun et al., CoNLL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/acl25-workshop-ingestion/2025.conll-1.11.pdf