@inproceedings{li-chan-2019-big,
    title = "Big Bidirectional Insertion Representations for Documents",
    author = "Li, Lala  and
      Chan, William",
    editor = "Birch, Alexandra  and
      Finch, Andrew  and
      Hayashi, Hiroaki  and
      Konstas, Ioannis  and
      Luong, Thang  and
      Neubig, Graham  and
      Oda, Yusuke  and
      Sudoh, Katsuhito",
    booktitle = "Proceedings of the 3rd Workshop on Neural Generation and Translation",
    month = nov,
    year = "2019",
    address = "Hong Kong",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/D19-5620/",
    doi = "10.18653/v1/D19-5620",
    pages = "194--198",
    abstract = "The Insertion Transformer is well suited for long form text generation due to its parallel generation capabilities, requiring $O(\log_2 n)$ generation steps to generate $n$ tokens. However, modeling long sequences is difficult, as there is more ambiguity captured in the attention mechanism. This work proposes the Big Bidirectional Insertion Representations for Documents (Big BIRD), an insertion-based model for document-level translation tasks. We scale up the insertion-based models to long form documents. Our key contribution is introducing sentence alignment via sentence-positional embeddings between the source and target document. We show an improvement of +4.3 BLEU on the WMT{'}19 English-{\ensuremath{>}}German document-level translation task compared with the Insertion Transformer baseline."
}Markdown (Informal)
[Big Bidirectional Insertion Representations for Documents](https://preview.aclanthology.org/ingest-emnlp/D19-5620/) (Li & Chan, NGT 2019)
ACL