Following Length Constraints in Instructions

Weizhe Yuan, Ilia Kulikov, Ping Yu, Kyunghyun Cho, Sainbayar Sukhbaatar, Jason E Weston, Jing Xu


Abstract
Aligned instruction following models can better fulfill user requests than their unaligned counterparts. However, it has been shown that there is a length bias in evaluation of such models, and that training algorithms tend to exploit this bias by learning longer responses. In this work we show how to train models that can be controlled at inference time with instructions containing desired length constraints. Such models are superior in length instructed evaluations, outperforming standard instruction following models such as GPT4, Llama 3 and Mixtral.
Anthology ID:
2025.emnlp-main.1233
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
24243–24254
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1233/
DOI:
Bibkey:
Cite (ACL):
Weizhe Yuan, Ilia Kulikov, Ping Yu, Kyunghyun Cho, Sainbayar Sukhbaatar, Jason E Weston, and Jing Xu. 2025. Following Length Constraints in Instructions. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 24243–24254, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Following Length Constraints in Instructions (Yuan et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1233.pdf
Checklist:
 2025.emnlp-main.1233.checklist.pdf