An Iterative Knowledge Transfer NMT System for WMT20 News Translation Task

Jiwan Kim, Soyoon Park, Sangha Kim, Yoonjung Choi


Abstract
This paper describes our submission to the WMT20 news translation shared task in English to Japanese direction. Our main approach is based on transferring knowledge of domain and linguistic characteristics by pre-training the encoder-decoder model with large amount of in-domain monolingual data through unsupervised and supervised prediction task. We then fine-tune the model with parallel data and in-domain synthetic data, generated with iterative back-translation. For additional gain, we generate final results with an ensemble model and re-rank them with averaged models and language models. Through these methods, we achieve +5.42 BLEU score compare to the baseline model.
Anthology ID:
2020.wmt-1.11
Volume:
Proceedings of the Fifth Conference on Machine Translation
Month:
November
Year:
2020
Address:
Online
Venue:
WMT
SIG:
SIGMT
Publisher:
Association for Computational Linguistics
Note:
Pages:
139–144
Language:
URL:
https://aclanthology.org/2020.wmt-1.11
DOI:
Bibkey:
Cite (ACL):
Jiwan Kim, Soyoon Park, Sangha Kim, and Yoonjung Choi. 2020. An Iterative Knowledge Transfer NMT System for WMT20 News Translation Task. In Proceedings of the Fifth Conference on Machine Translation, pages 139–144, Online. Association for Computational Linguistics.
Cite (Informal):
An Iterative Knowledge Transfer NMT System for WMT20 News Translation Task (Kim et al., WMT 2020)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-script-update/2020.wmt-1.11.pdf
Video:
 https://slideslive.com/38939599