@inproceedings{tan-etal-2019-hierarchical,
title = "Hierarchical Modeling of Global Context for Document-Level Neural Machine Translation",
author = "Tan, Xin and
Zhang, Longyin and
Xiong, Deyi and
Zhou, Guodong",
editor = "Inui, Kentaro and
Jiang, Jing and
Ng, Vincent and
Wan, Xiaojun",
booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP)",
month = nov,
year = "2019",
address = "Hong Kong, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest_wac_2008/D19-1168/",
doi = "10.18653/v1/D19-1168",
pages = "1576--1585",
abstract = "Document-level machine translation (MT) remains challenging due to the difficulty in efficiently using document context for translation. In this paper, we propose a hierarchical model to learn the global context for document-level neural machine translation (NMT). This is done through a sentence encoder to capture intra-sentence dependencies and a document encoder to model document-level inter-sentence consistency and coherence. With this hierarchical architecture, we feedback the extracted global document context to each word in a top-down fashion to distinguish different translations of a word according to its specific surrounding context. In addition, since large-scale in-domain document-level parallel corpora are usually unavailable, we use a two-step training strategy to take advantage of a large-scale corpus with out-of-domain parallel sentence pairs and a small-scale corpus with in-domain parallel document pairs to achieve the domain adaptability. Experimental results on several benchmark corpora show that our proposed model can significantly improve document-level translation performance over several strong NMT baselines."
}
Markdown (Informal)
[Hierarchical Modeling of Global Context for Document-Level Neural Machine Translation](https://preview.aclanthology.org/ingest_wac_2008/D19-1168/) (Tan et al., EMNLP-IJCNLP 2019)
ACL