Can Constructions “SCAN” Compositionality ?

Ganesh Katrapati, Manish Shrivastava


Abstract
Sequence to Sequence models struggle at compositionality and systematic generalisation even while they excel at many other tasks.We attribute this limitation to their failure to internalise constructions—conventionalised form–meaning pairings that license productive recombination. Building on these insights, we introduce an unsupervised procedure for mining pseudo-constructions: variable-slot templates automatically extracted from training data. When applied to the SCAN dataset, ourmethod yields large gains out-of-distribution splits: accuracy rises to 47.8% on ADD JUMP and to 20.3% on AROUND RIGHT without any architectural changes or additional supervision. The model also attains competitive performance with ≤ 40% of the original training data, demonstrating strong data efficiency. Our findings highlight the promise of construction-aware preprocessing as an alternative to heavy architectural or training-regime interventions.
Anthology ID:
2025.cxgsnlp-1.17
Volume:
Proceedings of the Second International Workshop on Construction Grammars and NLP
Month:
September
Year:
2025
Address:
Düsseldorf, Germany
Editors:
Claire Bonial, Melissa Torgbi, Leonie Weissweiler, Austin Blodgett, Katrien Beuls, Paul Van Eecke, Harish Tayyar Madabushi
Venues:
CxGsNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
165–171
Language:
URL:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.cxgsnlp-1.17/
DOI:
Bibkey:
Cite (ACL):
Ganesh Katrapati and Manish Shrivastava. 2025. Can Constructions “SCAN” Compositionality ?. In Proceedings of the Second International Workshop on Construction Grammars and NLP, pages 165–171, Düsseldorf, Germany. Association for Computational Linguistics.
Cite (Informal):
Can Constructions “SCAN” Compositionality ? (Katrapati & Shrivastava, CxGsNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.cxgsnlp-1.17.pdf