@inproceedings{wang-etal-2023-hybrid,
title = "Hybrid-Regressive Paradigm for Accurate and Speed-Robust Neural Machine Translation",
author = "Wang, Qiang and
Hu, Xinhui and
Chen, Ming",
editor = "Rogers, Anna and
Boyd-Graber, Jordan and
Okazaki, Naoaki",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2023",
month = jul,
year = "2023",
address = "Toronto, Canada",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.findings-acl.367/",
doi = "10.18653/v1/2023.findings-acl.367",
pages = "5931--5945",
abstract = "This work empirically confirms that non-autoregressive translation (NAT) is less robust in decoding batch size and hardware settings than autoregressive translation (AT). To address this issue, we demonstrate that prompting a small number of AT predictions can significantly reduce the performance gap between AT and NAT through synthetic experiments. Following this line, we propose hybrid-regressive translation (HRT), a two-stage translation prototype that combines the strengths of AT and NAT. Specifically, HRT first generates discontinuous sequences via autoregression (e.g., make a prediction for every $k$ tokens, $k>1$) and then fills in all previously skipped tokens at once in a non-autoregressive manner. Experiments on five translation tasks show that HRT achieves comparable translation quality with AT while having at least 1.5x faster inference regardless of batch size and device. Additionally, HRT successfully inherits the sound characteristics of AT in the deep-encoder-shallow-decoder architecture, allowing for further speedup without BLEU loss."
}
Markdown (Informal)
[Hybrid-Regressive Paradigm for Accurate and Speed-Robust Neural Machine Translation](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.findings-acl.367/) (Wang et al., Findings 2023)
ACL