LLMs Learn Constructions That Humans Do Not Know

Jonathan Dunn, Mai Mohamed Eida


Abstract
This paper investigates false positive constructions: grammatical structures which an LLM hallucinates as distinct constructions but which human introspection does not support. Both a behavioural probing task using contextual embeddings and a meta-linguistic probing task using prompts are included, allowing us to distinguish between implicit and explicit linguistic knowledge. Both methods reveal that models do indeed hallucinate constructions. We then simulate hypothesis testing to determine what would have happened if a linguist had falsely hypothesized that these hallucinated constructions do exist. The high accuracy obtained shows that such false hypotheses would have been overwhelmingly confirmed. This suggests that construction probing methods suffer from a confirmation bias and raises the issue of what unknown and incorrect syntactic knowledge these models also possess.
Anthology ID:
2025.cxgsnlp-1.2
Volume:
Proceedings of the Second International Workshop on Construction Grammars and NLP
Month:
September
Year:
2025
Address:
Düsseldorf, Germany
Editors:
Claire Bonial, Melissa Torgbi, Leonie Weissweiler, Austin Blodgett, Katrien Beuls, Paul Van Eecke, Harish Tayyar Madabushi
Venues:
CxGsNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
13–23
Language:
URL:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.cxgsnlp-1.2/
DOI:
Bibkey:
Cite (ACL):
Jonathan Dunn and Mai Mohamed Eida. 2025. LLMs Learn Constructions That Humans Do Not Know. In Proceedings of the Second International Workshop on Construction Grammars and NLP, pages 13–23, Düsseldorf, Germany. Association for Computational Linguistics.
Cite (Informal):
LLMs Learn Constructions That Humans Do Not Know (Dunn & Eida, CxGsNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.cxgsnlp-1.2.pdf