@inproceedings{alkhouli-etal-2018-alignment,
    title = "On The Alignment Problem In Multi-Head Attention-Based Neural Machine Translation",
    author = "Alkhouli, Tamer  and
      Bretschner, Gabriel  and
      Ney, Hermann",
    editor = "Bojar, Ond{\v{r}}ej  and
      Chatterjee, Rajen  and
      Federmann, Christian  and
      Fishel, Mark  and
      Graham, Yvette  and
      Haddow, Barry  and
      Huck, Matthias  and
      Yepes, Antonio Jimeno  and
      Koehn, Philipp  and
      Monz, Christof  and
      Negri, Matteo  and
      N{\'e}v{\'e}ol, Aur{\'e}lie  and
      Neves, Mariana  and
      Post, Matt  and
      Specia, Lucia  and
      Turchi, Marco  and
      Verspoor, Karin",
    booktitle = "Proceedings of the Third Conference on Machine Translation: Research Papers",
    month = oct,
    year = "2018",
    address = "Brussels, Belgium",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/iwcs-25-ingestion/W18-6318/",
    doi = "10.18653/v1/W18-6318",
    pages = "177--185",
    abstract = "This work investigates the alignment problem in state-of-the-art multi-head attention models based on the transformer architecture. We demonstrate that alignment extraction in transformer models can be improved by augmenting an additional alignment head to the multi-head source-to-target attention component. This is used to compute sharper attention weights. We describe how to use the alignment head to achieve competitive performance. To study the effect of adding the alignment head, we simulate a dictionary-guided translation task, where the user wants to guide translation using pre-defined dictionary entries. Using the proposed approach, we achieve up to 3.8{\%} BLEU improvement when using the dictionary, in comparison to 2.4{\%} BLEU in the baseline case. We also propose alignment pruning to speed up decoding in alignment-based neural machine translation (ANMT), which speeds up translation by a factor of 1.8 without loss in translation performance. We carry out experiments on the shared WMT 2016 English{\textrightarrow}Romanian news task and the BOLT Chinese{\textrightarrow}English discussion forum task."
}Markdown (Informal)
[On The Alignment Problem In Multi-Head Attention-Based Neural Machine Translation](https://preview.aclanthology.org/iwcs-25-ingestion/W18-6318/) (Alkhouli et al., WMT 2018)
ACL