BVSLP: Machine Translation Using Linguistic Embellishments for IndicMT Shared Task 2025

Nisheeth Joshi, Palak Arora, Anju Krishnia, Riya Lonchenpa, Mhasilenuo Vizo


Abstract
This paper describes our submission to the Indic MT 2025 shared task, where we trained machine translation systems for five low-resource language pairs: English–Manipuri, Manipuri–English, English–Bodo, English–Assamese, and Assamese–English. To address the challenge of out-of-vocabulary errors, we introduced a Named Entity Translation module that automatically identified named entities and either translated or transliterated them into the target language. The augmented corpus produced by this module was used to fine-tune a Transformer-based neural machine translation system. Our approach, termed HEMANT (Highly Efficient Machine-Assisted Natural Translation), demonstrated consistent improvements, particularly in reducing named entity errors and improving fluency for Assamese–English and Manipuri–English. Official shared task evaluation results show that the system achieved competitive performance across all five language pairs, underscoring the effectiveness of linguistically informed preprocessing for low-resource Indic MT.
Anthology ID:
2025.wmt-1.105
Volume:
Proceedings of the Tenth Conference on Machine Translation
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Barry Haddow, Tom Kocmi, Philipp Koehn, Christof Monz
Venue:
WMT
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
1265–1270
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.105/
DOI:
Bibkey:
Cite (ACL):
Nisheeth Joshi, Palak Arora, Anju Krishnia, Riya Lonchenpa, and Mhasilenuo Vizo. 2025. BVSLP: Machine Translation Using Linguistic Embellishments for IndicMT Shared Task 2025. In Proceedings of the Tenth Conference on Machine Translation, pages 1265–1270, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
BVSLP: Machine Translation Using Linguistic Embellishments for IndicMT Shared Task 2025 (Joshi et al., WMT 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.105.pdf