CASIA’s System for IWSLT 2020 Open Domain Translation

Qian Wang, Yuchen Liu, Cong Ma, Yu Lu, Yining Wang, Long Zhou, Yang Zhao, Jiajun Zhang, Chengqing Zong


Abstract
This paper describes the CASIA’s system for the IWSLT 2020 open domain translation task. This year we participate in both Chinese→Japanese and Japanese→Chinese translation tasks. Our system is neural machine translation system based on Transformer model. We augment the training data with knowledge distillation and back translation to improve the translation performance. Domain data classification and weighted domain model ensemble are introduced to generate the final translation result. We compare and analyze the performance on development data with different model settings and different data processing techniques.
Anthology ID:
2020.iwslt-1.15
Volume:
Proceedings of the 17th International Conference on Spoken Language Translation
Month:
July
Year:
2020
Address:
Online
Venue:
IWSLT
SIG:
SIGSLT
Publisher:
Association for Computational Linguistics
Note:
Pages:
130–139
Language:
URL:
https://aclanthology.org/2020.iwslt-1.15
DOI:
10.18653/v1/2020.iwslt-1.15
Bibkey:
Cite (ACL):
Qian Wang, Yuchen Liu, Cong Ma, Yu Lu, Yining Wang, Long Zhou, Yang Zhao, Jiajun Zhang, and Chengqing Zong. 2020. CASIA’s System for IWSLT 2020 Open Domain Translation. In Proceedings of the 17th International Conference on Spoken Language Translation, pages 130–139, Online. Association for Computational Linguistics.
Cite (Informal):
CASIA’s System for IWSLT 2020 Open Domain Translation (Wang et al., IWSLT 2020)
Copy Citation:
PDF:
https://preview.aclanthology.org/emnlp-22-attachments/2020.iwslt-1.15.pdf
Video:
 http://slideslive.com/38929589