Transfer learning in low-resourced MT: An empirical study
Mahata Sainik, Saha Dipanjan, Das Dipankar, Bandyopadhyay Sivaji
Abstract
Translation systems rely on a large and goodquality parallel corpus for producing reliable translations. However, obtaining such a corpus for low-resourced languages is a challenge. New research has shown that transfer learning can mitigate this issue by augmenting lowresourced MT systems with high-resourced ones. In this work, we explore two types of transfer learning techniques, namely, crosslingual transfer learning and multilingual training, both with information augmentation, to examine the degree of performance improvement following the augmentation. Furthermore, we use languages of the same family (Romanic, in our case), to investigate the role of the shared linguistic property, in producing dependable translations.- Anthology ID:
- 2023.icon-1.63
- Volume:
- Proceedings of the 20th International Conference on Natural Language Processing (ICON)
- Month:
- December
- Year:
- 2023
- Address:
- Goa University, Goa, India
- Editors:
- D. Pawar Jyoti, Lalitha Devi Sobha
- Venue:
- ICON
- SIG:
- SIGLEX
- Publisher:
- NLP Association of India (NLPAI)
- Note:
- Pages:
- 646–650
- Language:
- URL:
- https://aclanthology.org/2023.icon-1.63
- DOI:
- Cite (ACL):
- Mahata Sainik, Saha Dipanjan, Das Dipankar, and Bandyopadhyay Sivaji. 2023. Transfer learning in low-resourced MT: An empirical study. In Proceedings of the 20th International Conference on Natural Language Processing (ICON), pages 646–650, Goa University, Goa, India. NLP Association of India (NLPAI).
- Cite (Informal):
- Transfer learning in low-resourced MT: An empirical study (Sainik et al., ICON 2023)
- PDF:
- https://preview.aclanthology.org/fix-volume-bibkeys/2023.icon-1.63.pdf