@inproceedings{dat-etal-2025-discrete,
title = "Discrete Diffusion Language Model for Efficient Text Summarization",
author = "Dat, Do Huu and
Do, Duc Anh and
Luu, Anh Tuan and
Buntine, Wray",
editor = "Chiruzzo, Luis and
Ritter, Alan and
Wang, Lu",
booktitle = "Findings of the Association for Computational Linguistics: NAACL 2025",
month = apr,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.352/",
pages = "6278--6290",
ISBN = "979-8-89176-195-7",
abstract = "While diffusion models excel at conditionally generating high-quality images, prior works in discrete diffusion models were not evaluated on conditional long-text generation. This work addresses the limitations of prior discrete diffusion models for conditional long-text generation, particularly in the long abstractive summarization task. Despite faster decoding speeds compared to autoregressive methods, previous discrete diffusion models failed on the abstractive summarization task due to the incompatibility between the backbone architectures and the random noising process. To overcome these challenges, we introduce a novel semantic-aware noising process that enables Transformer backbones to handle long sequences effectively. Additionally, we propose CrossMamba, an adaptation of the Mamba model to the encoder-decoder paradigm, which integrates seamlessly with the random absorbing noising process. Our approaches outperform existing discrete diffusion models on three benchmark summarization datasets: Gigaword, CNN/DailyMail, and Arxiv, while also achieving much faster inference speed compared to autoregressive models."
}
Markdown (Informal)
[Discrete Diffusion Language Model for Efficient Text Summarization](https://preview.aclanthology.org/fix-sig-urls/2025.findings-naacl.352/) (Dat et al., Findings 2025)
ACL