Lucy Park
2021
Unsupervised Neural Machine Translation for Low-Resource Domains via Meta-Learning
Cheonbok Park
|
Yunwon Tae
|
TaeHee Kim
|
Soyoung Yang
|
Mohammad Azam Khan
|
Lucy Park
|
Jaegul Choo
Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
Unsupervised machine translation, which utilizes unpaired monolingual corpora as training data, has achieved comparable performance against supervised machine translation. However, it still suffers from data-scarce domains. To address this issue, this paper presents a novel meta-learning algorithm for unsupervised neural machine translation (UNMT) that trains the model to adapt to another domain by utilizing only a small amount of training data. We assume that domain-general knowledge is a significant factor in handling data-scarce domains. Hence, we extend the meta-learning algorithm, which utilizes knowledge learned from high-resource domains, to boost the performance of low-resource UNMT. Our model surpasses a transfer learning-based approach by up to 2-3 BLEU scores. Extensive experimental results show that our proposed algorithm is pertinent for fast adaptation and consistently outperforms other baselines.
Search
Co-authors
- Cheonbok Park 1
- Yunwon Tae 1
- Taehee Kim 1
- Soyoung Yang 1
- Mohammad Azam Khan 1
- show all...