Using Encipherment to Isolate Conditions for the Successful Fine-tuning of Massively Multilingual Translation Models

Carter Louchheim, Denis Sotnichenko, Yukina Yamaguchi, Mark Hopkins


Abstract
When fine-tuning massively multilingual translation models for low-resource languages, practitioners often include auxiliary languages to improve performance, but factors determining successful auxiliary language selection remain unclear. This paper investigates whether syntactic similarity or lexical overlap is more important for effective multilingual fine-tuning. We use encipherment to create controlled experimental conditions that disentangle these confounded factors, generating novel languages with identical syntax but no lexical overlap, and conversely, languages that preserve lexical overlap. Through extensive NLLB-200 fine-tuning experiments across Europarl and AmericasNLP datasets, we demonstrate that lexical overlap is the dominant factor. Syntactically identical auxiliary languages provide negligible benefits ( 1.0 ChrF), while languages with significant lexical overlap provide substantial improvements ( 5.0 ChrF), with effectiveness strongly correlated to KL-divergence between token distributions (r = -0.47, p .001). Our findings provide clear guidance: when selecting auxiliary languages for multilingual fine-tuning, prioritize lexical overlap over syntactic similarity.
Anthology ID:
2025.wmt-1.14
Volume:
Proceedings of the Tenth Conference on Machine Translation
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Barry Haddow, Tom Kocmi, Philipp Koehn, Christof Monz
Venue:
WMT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
241–252
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.14/
DOI:
Bibkey:
Cite (ACL):
Carter Louchheim, Denis Sotnichenko, Yukina Yamaguchi, and Mark Hopkins. 2025. Using Encipherment to Isolate Conditions for the Successful Fine-tuning of Massively Multilingual Translation Models. In Proceedings of the Tenth Conference on Machine Translation, pages 241–252, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Using Encipherment to Isolate Conditions for the Successful Fine-tuning of Massively Multilingual Translation Models (Louchheim et al., WMT 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.14.pdf