@inproceedings{ding-etal-2020-self,
title = "Self-Attention with Cross-Lingual Position Representation",
author = "Ding, Liang and
Wang, Longyue and
Tao, Dacheng",
editor = "Jurafsky, Dan and
Chai, Joyce and
Schluter, Natalie and
Tetreault, Joel",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/Add-Cong-Liu-Florida-Atlantic-University-author-id/2020.acl-main.153/",
doi = "10.18653/v1/2020.acl-main.153",
pages = "1679--1685",
abstract = "Position encoding (PE), an essential part of self-attention networks (SANs), is used to preserve the word order information for natural language processing tasks, generating fixed position indices for input sequences. However, in cross-lingual scenarios, machine translation, the PEs of source and target sentences are modeled independently. Due to word order divergences in different languages, modeling the cross-lingual positional relationships might help SANs tackle this problem. In this paper, we augment SANs with \textit{cross-lingual position representations} to model the bilingually aware latent structure for the input sentence. Specifically, we utilize bracketing transduction grammar (BTG)-based reordering information to encourage SANs to learn bilingual diagonal alignments. Experimental results on WMT`14 English$\Rightarrow$German, WAT`17 Japanese$\Rightarrow$English, and WMT`17 Chinese$\Leftrightarrow$English translation tasks demonstrate that our approach significantly and consistently improves translation quality over strong baselines. Extensive analyses confirm that the performance gains come from the cross-lingual information."
}
Markdown (Informal)
[Self-Attention with Cross-Lingual Position Representation](https://preview.aclanthology.org/Add-Cong-Liu-Florida-Atlantic-University-author-id/2020.acl-main.153/) (Ding et al., ACL 2020)
ACL