@inproceedings{yin-etal-2024-empirical,
title = "Empirical Prior for Text Autoencoders",
author = "Yin, Yongjing and
Gao, Wenyang and
Wu, Haodong and
Yan, Jianhao and
Zhang, Yue",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.findings-emnlp.796/",
doi = "10.18653/v1/2024.findings-emnlp.796",
pages = "13628--13640",
abstract = "This paper explores the application of Variational Autoencoders (VAE) in text generation, focusing on overcoming challenges like posterior collapse and the limitations of simplistic prior distributions. We investigate a transition from VAE to text autoencoders (AE), which model a compact latent space and preserves the capability of the language model itself. Our method involves layer-wise latent vectors regularized by orthogonal constraints to encourage distinct semantic spaces. In particular, we estimate an empirical prior online from the learned latent vectors to support sampling during generation like VAE. Experimental results on standard benchmarks demonstrate that the autoencoders generate higher quality and more diverse text than the VAE-based Transformer baselines, offering an effective alternative for generative language modeling."
}
Markdown (Informal)
[Empirical Prior for Text Autoencoders](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.findings-emnlp.796/) (Yin et al., Findings 2024)
ACL
- Yongjing Yin, Wenyang Gao, Haodong Wu, Jianhao Yan, and Yue Zhang. 2024. Empirical Prior for Text Autoencoders. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 13628–13640, Miami, Florida, USA. Association for Computational Linguistics.