Abstract
In this manuscript, we (team name is NLPRL) describe systems description that was submitted to the translation shared tasks at WAT 2020. We describe our model as transformer based NMT by using byte-level based BPE (BBPE). We used the OdiEnCorp 2.0 parallel corpus provided by the shared task organizer where the training, validation, and test data contain 69370, 13544, and 14344 lines of parallel sentences, respectively. The evaluation results show the BLEU score of English-to-Oria below the Organizer (1.34) and Oria-to-English direction shows above the Organizer (11.33).- Anthology ID:
- 2020.wat-1.13
- Volume:
- Proceedings of the 7th Workshop on Asian Translation
- Month:
- December
- Year:
- 2020
- Address:
- Suzhou, China
- Editors:
- Toshiaki Nakazawa, Hideki Nakayama, Chenchen Ding, Raj Dabre, Anoop Kunchukuttan, Win Pa Pa, Ondřej Bojar, Shantipriya Parida, Isao Goto, Hidaya Mino, Hiroshi Manabe, Katsuhito Sudoh, Sadao Kurohashi, Pushpak Bhattacharyya
- Venue:
- WAT
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 118–121
- Language:
- URL:
- https://aclanthology.org/2020.wat-1.13
- DOI:
- Cite (ACL):
- Rupjyoti Baruah and Rajesh Kumar Mundotiya. 2020. NLPRL Odia-English: Indic Language Neural Machine Translation System. In Proceedings of the 7th Workshop on Asian Translation, pages 118–121, Suzhou, China. Association for Computational Linguistics.
- Cite (Informal):
- NLPRL Odia-English: Indic Language Neural Machine Translation System (Baruah & Mundotiya, WAT 2020)
- PDF:
- https://preview.aclanthology.org/emnlp-22-attachments/2020.wat-1.13.pdf