BVSLP: Machine Translation Using Linguistic Embellishments for IndicMT Shared Task 2025
Nisheeth Joshi, Palak Arora, Anju Krishnia, Riya Lonchenpa, Mhasilenuo Vizo
Abstract
This paper describes our submission to the Indic MT 2025 shared task, where we trained machine translation systems for five low-resource language pairs: English–Manipuri, Manipuri–English, English–Bodo, English–Assamese, and Assamese–English. To address the challenge of out-of-vocabulary errors, we introduced a Named Entity Translation module that automatically identified named entities and either translated or transliterated them into the target language. The augmented corpus produced by this module was used to fine-tune a Transformer-based neural machine translation system. Our approach, termed HEMANT (Highly Efficient Machine-Assisted Natural Translation), demonstrated consistent improvements, particularly in reducing named entity errors and improving fluency for Assamese–English and Manipuri–English. Official shared task evaluation results show that the system achieved competitive performance across all five language pairs, underscoring the effectiveness of linguistically informed preprocessing for low-resource Indic MT.- Anthology ID:
- 2025.wmt-1.105
- Volume:
- Proceedings of the Tenth Conference on Machine Translation
- Month:
- November
- Year:
- 2025
- Address:
- Suzhou, China
- Editors:
- Barry Haddow, Tom Kocmi, Philipp Koehn, Christof Monz
- Venue:
- WMT
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 1265–1270
- Language:
- URL:
- https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.105/
- DOI:
- Cite (ACL):
- Nisheeth Joshi, Palak Arora, Anju Krishnia, Riya Lonchenpa, and Mhasilenuo Vizo. 2025. BVSLP: Machine Translation Using Linguistic Embellishments for IndicMT Shared Task 2025. In Proceedings of the Tenth Conference on Machine Translation, pages 1265–1270, Suzhou, China. Association for Computational Linguistics.
- Cite (Informal):
- BVSLP: Machine Translation Using Linguistic Embellishments for IndicMT Shared Task 2025 (Joshi et al., WMT 2025)
- PDF:
- https://preview.aclanthology.org/ingest-emnlp/2025.wmt-1.105.pdf