Enable Fast Sampling for Seq2Seq Text Diffusion

Pan Liu, Xiaohua Tian, Zhouhan Lin


Abstract
Diffusion models exhibit promising capacity for generating high-quality text. However, owing to the curved nature of generation path, they necessitate traversing numerous steps to guarantee the text quality. In this paper, we propose an efficient model FMSeq, which utilizes flow matching to straighten the generation path, thereby enabling fast sampling for diffusion-based seq2seq text generation. Specifically, we construct transport flow only on the target sequences to adapt the diffusion-based model with flow matching. Furthermore, we explore different settings and identify target-parameterization, self-conditioning and time-difference as three effective techniques to improve the generation quality under a few steps. Experiments on four popular tasks demonstrate that FMSeq generates texts of comparable quality to the SOTA diffusion-based DiffuSeq in just 10 steps, achieving a 200-fold speedup.
Anthology ID:
2024.findings-emnlp.497
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2024
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
8495–8505
Language:
URL:
https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.497/
DOI:
10.18653/v1/2024.findings-emnlp.497
Bibkey:
Cite (ACL):
Pan Liu, Xiaohua Tian, and Zhouhan Lin. 2024. Enable Fast Sampling for Seq2Seq Text Diffusion. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 8495–8505, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
Enable Fast Sampling for Seq2Seq Text Diffusion (Liu et al., Findings 2024)
Copy Citation:
PDF:
https://preview.aclanthology.org/jlcl-multiple-ingestion/2024.findings-emnlp.497.pdf