Babies Learn to Look Ahead: Multi-Token Prediction in Small LMs

Ansar Aynetdinov, Alan Akbik


Abstract
Multi-token prediction (MTP) is an alternative training objective for language models that has recently been proposed as a potential improvement over traditional next-token prediction (NTP). Instead of training models to predict only the next token, as is standard, MTP trains them to predict the next k tokens at each step. While MTP was shown to improve downstream performance and sample efficiency in large language models (LLMs), smaller language models (SLMs) struggle with this objective. Recently, a curriculum-based approach was offered as a solution to this problem for models as small as 1.3B parameters by adjusting the difficulty of the training objective over time. In this work we investigate the viability of MTP curricula in a highly data- and parameter-constrained setting. Our experimental results show that even 130M-parameter models benefit from including the MTP task in the pre-training objective. These gains hold even under severe data constraints, as demonstrated on both zero-shot benchmarks and downstream tasks.
Anthology ID:
2025.babylm-main.41
Volume:
Proceedings of the First BabyLM Workshop
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Lucas Charpentier, Leshem Choshen, Ryan Cotterell, Mustafa Omer Gul, Michael Y. Hu, Jing Liu, Jaap Jumelet, Tal Linzen, Aaron Mueller, Candace Ross, Raj Sanjay Shah, Alex Warstadt, Ethan Gotlieb Wilcox, Adina Williams
Venue:
BabyLM
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
566–577
Language:
URL:
https://preview.aclanthology.org/setup/2025.babylm-main.41/
DOI:
Bibkey:
Cite (ACL):
Ansar Aynetdinov and Alan Akbik. 2025. Babies Learn to Look Ahead: Multi-Token Prediction in Small LMs. In Proceedings of the First BabyLM Workshop, pages 566–577, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Babies Learn to Look Ahead: Multi-Token Prediction in Small LMs (Aynetdinov & Akbik, BabyLM 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/setup/2025.babylm-main.41.pdf