RecombiText: Compositional Data Augmentation for Enhancing LLM Pre-Training Datasets in Low-Resource Scenarios
Alexander Tampier, Lukas Thoma, Loris Schoenegger, Benjamin Roth
Abstract
We introduce RecombiText Augmentation (RTA), a novel purely statistical NLP method for compositional data augmentation for data-efficient LLM pre-training in low-resource scenarios. RTA identifies lexically and semantically similar sentences within the corpus and generates synthetic sentence pairs from them while preserving underlying patterns from the corpus. We pre-train GPT-2 and RoBERTa language models on a domain-specific, low-resource corpus of 10 million words, with different proportions of augmented data. We compare our RTA-augmented model variants to a baseline model trained on the full original dataset. Zero-shot results show that the language models pre-trained on synthetic data improve in entity tracking, self-paced reading, and morphological generalization benchmarks. In other tasks, the performance is comparable to the baseline model. We demonstrate that it is possible to expand low-resource datasets by two- to four-fold without compromising benchmark performance, solely through statistical processing of the available data.- Anthology ID:
- 2025.babylm-main.40
- Volume:
- Proceedings of the First BabyLM Workshop
- Month:
- November
- Year:
- 2025
- Address:
- Suzhou, China
- Editors:
- Lucas Charpentier, Leshem Choshen, Ryan Cotterell, Mustafa Omer Gul, Michael Y. Hu, Jing Liu, Jaap Jumelet, Tal Linzen, Aaron Mueller, Candace Ross, Raj Sanjay Shah, Alex Warstadt, Ethan Gotlieb Wilcox, Adina Williams
- Venue:
- BabyLM
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 548–565
- Language:
- URL:
- https://preview.aclanthology.org/ingest-emnlp/2025.babylm-main.40/
- DOI:
- Cite (ACL):
- Alexander Tampier, Lukas Thoma, Loris Schoenegger, and Benjamin Roth. 2025. RecombiText: Compositional Data Augmentation for Enhancing LLM Pre-Training Datasets in Low-Resource Scenarios. In Proceedings of the First BabyLM Workshop, pages 548–565, Suzhou, China. Association for Computational Linguistics.
- Cite (Informal):
- RecombiText: Compositional Data Augmentation for Enhancing LLM Pre-Training Datasets in Low-Resource Scenarios (Tampier et al., BabyLM 2025)
- PDF:
- https://preview.aclanthology.org/ingest-emnlp/2025.babylm-main.40.pdf