@inproceedings{gumma-etal-2025-towards,
title = "Towards Inducing Long-Context Abilities in Multilingual Neural Machine Translation Models",
author = "Gumma, Varun and
Chitale, Pranjal A and
Bali, Kalika",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.naacl-long.366/",
pages = "7158--7170",
ISBN = "979-8-89176-189-6",
abstract = "Neural Machine Translation (NMT) models have traditionally used Sinusoidal Positional Embeddings (PEs), which often struggle to capture long-range dependencies and are inefficient for handling extended context or document-level translation tasks. This work addresses the challenge of transitioning pre-trained NMT models from absolute Sinusoidal PEs to Relative PEs, such as RoPE and ALiBi, without compromising performance. We demonstrate that parameter-efficient fine-tuning, using only a small amount of high-quality data, can successfully facilitate this transition. Experimental results indicate that switching from Sinusoidal to Relative PEs results in competitive translation quality on sentence-level evaluation benchmarks. Additionally, models trained with RoPE consistently outperform those using ALiBi and Sinusoidal PEs on document-level benchmarks across both string-based metrics and qualitative evaluations. Moreover, we find that a small amount of long-context data in a few languages is sufficient for cross-lingual length generalization, thereby inducing long-context capabilities."
}
Markdown (Informal)
[Towards Inducing Long-Context Abilities in Multilingual Neural Machine Translation Models](https://preview.aclanthology.org/fix-sig-urls/2025.naacl-long.366/) (Gumma et al., NAACL 2025)
ACL