Challenge Track: Divide and Translate: Parameter Isolation with Encoder Freezing for Low-Resource Indic NMT

Vaibhav Kanojia


Abstract
We present a direction-specialized neural machine translation framework for ultra-low-resource Indic and tribal languages, including Bhili, Gondi, Mundari, and Santali. Using the NLLB-600M backbone, we freeze the multilingual encoder and fine-tune direction-specific decoders to reduce negative transfer and improve morphological fidelity under severe data scarcity. Our system is trained with leakage-safe splits, bitext reversal augmentation, and memory-efficient mixed-precision optimization. On the official MMLoSo 2025 Kaggle benchmark, we achieve a public score of 171.4 and a private score of 161.1, demonstrating stable generalization in highly noisy low-resource conditions.
Anthology ID:
2025.mmloso-1.12
Volume:
Proceedings of the 1st Workshop on Multimodal Models for Low-Resource Contexts and Social Impact (MMLoSo 2025)
Month:
December
Year:
2025
Address:
Mumbai, India
Editors:
Ankita Shukla, Sandeep Kumar, Amrit Singh Bedi, Tanmoy Chakraborty
Venues:
MMLoSo | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
109–113
Language:
URL:
https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.mmloso-1.12/
DOI:
Bibkey:
Cite (ACL):
Vaibhav Kanojia. 2025. Challenge Track: Divide and Translate: Parameter Isolation with Encoder Freezing for Low-Resource Indic NMT. In Proceedings of the 1st Workshop on Multimodal Models for Low-Resource Contexts and Social Impact (MMLoSo 2025), pages 109–113, Mumbai, India. Association for Computational Linguistics.
Cite (Informal):
Challenge Track: Divide and Translate: Parameter Isolation with Encoder Freezing for Low-Resource Indic NMT (Kanojia, MMLoSo 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.mmloso-1.12.pdf