A Document-Level Neural Machine Translation Model with Dynamic Caching Guided by Theme-Rheme Information

Yiqi Tong, Jiangbin Zheng, Hongkang Zhu, Yidong Chen, Xiaodong Shi


Abstract
Research on document-level Neural Machine Translation (NMT) models has attracted increasing attention in recent years. Although the proposed works have proved that the inter-sentence information is helpful for improving the performance of the NMT models, what information should be regarded as context remains ambiguous. To solve this problem, we proposed a novel cache-based document-level NMT model which conducts dynamic caching guided by theme-rheme information. The experiments on NIST evaluation sets demonstrate that our proposed model achieves substantial improvements over the state-of-the-art baseline NMT models. As far as we know, we are the first to introduce theme-rheme theory into the field of machine translation.
Anthology ID:
2020.coling-main.388
Volume:
Proceedings of the 28th International Conference on Computational Linguistics
Month:
December
Year:
2020
Address:
Barcelona, Spain (Online)
Editors:
Donia Scott, Nuria Bel, Chengqing Zong
Venue:
COLING
SIG:
Publisher:
International Committee on Computational Linguistics
Note:
Pages:
4385–4395
Language:
URL:
https://aclanthology.org/2020.coling-main.388
DOI:
10.18653/v1/2020.coling-main.388
Bibkey:
Cite (ACL):
Yiqi Tong, Jiangbin Zheng, Hongkang Zhu, Yidong Chen, and Xiaodong Shi. 2020. A Document-Level Neural Machine Translation Model with Dynamic Caching Guided by Theme-Rheme Information. In Proceedings of the 28th International Conference on Computational Linguistics, pages 4385–4395, Barcelona, Spain (Online). International Committee on Computational Linguistics.
Cite (Informal):
A Document-Level Neural Machine Translation Model with Dynamic Caching Guided by Theme-Rheme Information (Tong et al., COLING 2020)
Copy Citation:
PDF:
https://preview.aclanthology.org/nschneid-patch-5/2020.coling-main.388.pdf
Code
 ydc/ctrd