@inproceedings{tu-etal-2020-engine,
title = "{ENGINE}: Energy-Based Inference Networks for Non-Autoregressive Machine Translation",
author = "Tu, Lifu and
Pang, Richard Yuanzhe and
Wiseman, Sam and
Gimpel, Kevin",
editor = "Jurafsky, Dan and
Chai, Joyce and
Schluter, Natalie and
Tetreault, Joel",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2020.acl-main.251/",
doi = "10.18653/v1/2020.acl-main.251",
pages = "2819--2826",
abstract = "We propose to train a non-autoregressive machine translation model to minimize the energy defined by a pretrained autoregressive model. In particular, we view our non-autoregressive translation system as an inference network (Tu and Gimpel, 2018) trained to minimize the autoregressive teacher energy. This contrasts with the popular approach of training a non-autoregressive model on a distilled corpus consisting of the beam-searched outputs of such a teacher model. Our approach, which we call ENGINE (ENerGy-based Inference NEtworks), achieves state-of-the-art non-autoregressive results on the IWSLT 2014 DE-EN and WMT 2016 RO-EN datasets, approaching the performance of autoregressive models."
}
Markdown (Informal)
[ENGINE: Energy-Based Inference Networks for Non-Autoregressive Machine Translation](https://preview.aclanthology.org/fix-sig-urls/2020.acl-main.251/) (Tu et al., ACL 2020)
ACL