BabyLM’s First Constructions: Causal interventions provide a signal of learning

Joshua Rozner, Leonie Weissweiler, Cory Shain


Abstract
Construction grammar posits that language learners acquire constructions (form-meaning pairings) from the statistics of their environment. Recent work supports this hypothesis by showing sensitivity to constructions in pretrained language models (PLMs), including one recent study (Rozner et al., 2025) demonstrating that constructions shape RoBERTa’s output distribution. However, models under study have generally been trained on developmentally implausible amounts of data, casting doubt on their relevance to human language learning. Here we use Rozner et al.’s methods to evaluate construction learning in masked language models from the 2024 BabyLM Challenge.Our results show that even when trained on developmentally plausible quantities of data, models learn diverse constructions, even hard cases that are superficially indistinguishable.We further find correlational evidence that constructional performance may be functionally relevant: models that better represent constructions perform better on the BabyLM benchmarks.
Anthology ID:
2025.emnlp-main.113
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2237–2249
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.113/
DOI:
Bibkey:
Cite (ACL):
Joshua Rozner, Leonie Weissweiler, and Cory Shain. 2025. BabyLM’s First Constructions: Causal interventions provide a signal of learning. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 2237–2249, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
BabyLM’s First Constructions: Causal interventions provide a signal of learning (Rozner et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.113.pdf
Checklist:
 2025.emnlp-main.113.checklist.pdf