Using Encipherment to Isolate Conditions for the Successful Fine-tuning of Massively Multilingual Translation Models
Carter Louchheim, Denis Sotnichenko, Yukina Yamaguchi, Mark Hopkins
Abstract
When fine-tuning massively multilingual translation models for low-resource languages, practitioners often include auxiliary languages to improve performance, but factors determining successful auxiliary language selection remain unclear. This paper investigates whether syntactic similarity or lexical overlap is more important for effective multilingual fine-tuning. We use encipherment to create controlled experimental conditions that disentangle these confounded factors, generating novel languages with identical syntax but no lexical overlap, and conversely, languages that preserve lexical overlap. Through extensive NLLB-200 fine-tuning experiments across Europarl and AmericasNLP datasets, we demonstrate that lexical overlap is the dominant factor. Syntactically identical auxiliary languages provide negligible benefits ( 1.0 ChrF), while languages with significant lexical overlap provide substantial improvements ( 5.0 ChrF), with effectiveness strongly correlated to KL-divergence between token distributions (r = -0.47, p .001). Our findings provide clear guidance: when selecting auxiliary languages for multilingual fine-tuning, prioritize lexical overlap over syntactic similarity.- Anthology ID:
- 2025.wmt-1.14
- Volume:
- Proceedings of the Tenth Conference on Machine Translation
- Month:
- November
- Year:
- 2025
- Address:
- Suzhou, China
- Editors:
- Barry Haddow, Tom Kocmi, Philipp Koehn, Christof Monz
- Venue:
- WMT
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 241–252
- Language:
- URL:
- https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.14/
- DOI:
- Cite (ACL):
- Carter Louchheim, Denis Sotnichenko, Yukina Yamaguchi, and Mark Hopkins. 2025. Using Encipherment to Isolate Conditions for the Successful Fine-tuning of Massively Multilingual Translation Models. In Proceedings of the Tenth Conference on Machine Translation, pages 241–252, Suzhou, China. Association for Computational Linguistics.
- Cite (Informal):
- Using Encipherment to Isolate Conditions for the Successful Fine-tuning of Massively Multilingual Translation Models (Louchheim et al., WMT 2025)
- PDF:
- https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.14.pdf