Construction-Grammar Informed Parameter Efficient Fine-Tuning for Language Models

Prasanth


Abstract
Large language models excel at statistical pattern recognition but may lack explicit understanding of constructional form-meaning correspondences that characterize human grammatical competence. This paper presents Construction-Aware LoRA (CA-LoRA), a parameter-efficient fine-tuning method that incorporates constructional templates through specialized loss functions and targeted parameter updates. We focus on five major English construction types: ditransitive, caused-motion, resultative, way-construction, and conative. Evaluation on BLiMP, CoLA, and SyntaxGym shows selective improvements: frequent patterns like ditransitive and caused-motion show improvements of approximately 3.5 percentage points, while semi-productive constructions show minimal benefits (1.2 points). Overall performance improves by 1.8% on BLiMP and 1.6% on SyntaxGym, while maintaining competitive performance on general NLP tasks. Our approach requires only 1.72% of trainable parameters and reduces training time by 67% compared to full fine-tuning. This work demonstrates that explicit constructional knowledge can be selectively integrated into neural language models, with effectiveness dependent on construction frequency and structural regularity.
Anthology ID:
2025.cxgsnlp-1.4
Volume:
Proceedings of the Second International Workshop on Construction Grammars and NLP
Month:
September
Year:
2025
Address:
Düsseldorf, Germany
Editors:
Claire Bonial, Melissa Torgbi, Leonie Weissweiler, Austin Blodgett, Katrien Beuls, Paul Van Eecke, Harish Tayyar Madabushi
Venues:
CxGsNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
34–40
Language:
URL:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.cxgsnlp-1.4/
DOI:
Bibkey:
Cite (ACL):
Prasanth. 2025. Construction-Grammar Informed Parameter Efficient Fine-Tuning for Language Models. In Proceedings of the Second International Workshop on Construction Grammars and NLP, pages 34–40, Düsseldorf, Germany. Association for Computational Linguistics.
Cite (Informal):
Construction-Grammar Informed Parameter Efficient Fine-Tuning for Language Models (Prasanth, CxGsNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/iwcs-25-ingestion/2025.cxgsnlp-1.4.pdf