Chao Bian
2018
Neural Machine Translation with Decoding History Enhanced Attention
Mingxuan Wang
|
Jun Xie
|
Zhixing Tan
|
Jinsong Su
|
Deyi Xiong
|
Chao Bian
Proceedings of the 27th International Conference on Computational Linguistics
Neural machine translation with source-side attention have achieved remarkable performance. however, there has been little work exploring to attend to the target-side which can potentially enhance the memory capbility of NMT. We reformulate a Decoding History Enhanced Attention mechanism (DHEA) to render NMT model better at selecting both source-side and target-side information. DHA enables dynamic control of the ratios at which source and target contexts contribute to the generation of target words, offering a way to weakly induce structure relations among both source and target tokens. It also allows training errors to be directly back-propagated through short-cut connections and effectively alleviates the gradient vanishing problem. The empirical study on Chinese-English translation shows that our model with proper configuration can improve by 0:9 BLEU upon Transformer and the best reported results in the dataset. On WMT14 English-German task and a larger WMT14 English-French task, our model achieves comparable results with the state-of-the-art.
Tencent Neural Machine Translation Systems for WMT18
Mingxuan Wang
|
Li Gong
|
Wenhuan Zhu
|
Jun Xie
|
Chao Bian
Proceedings of the Third Conference on Machine Translation: Shared Task Papers
We participated in the WMT 2018 shared news translation task on English↔Chinese language pair. Our systems are based on attentional sequence-to-sequence models with some form of recursion and self-attention. Some data augmentation methods are also introduced to improve the translation performance. The best translation result is obtained with ensemble and reranking techniques. Our Chinese→English system achieved the highest cased BLEU score among all 16 submitted systems, and our English→Chinese system ranked the third out of 18 submitted systems.
Search
Co-authors
- Mingxuan Wang 2
- Jun Xie 2
- Zhixing Tan 1
- Jinsong Su 1
- Deyi Xiong 1
- show all...