@inproceedings{bao-etal-2020-plato,
title = "{PLATO}: Pre-trained Dialogue Generation Model with Discrete Latent Variable",
author = "Bao, Siqi and
He, Huang and
Wang, Fan and
Wu, Hua and
Wang, Haifeng",
editor = "Jurafsky, Dan and
Chai, Joyce and
Schluter, Natalie and
Tetreault, Joel",
booktitle = "Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics",
month = jul,
year = "2020",
address = "Online",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2020.acl-main.9/",
doi = "10.18653/v1/2020.acl-main.9",
pages = "85--96",
abstract = "Pre-training models have been proved effective for a wide range of natural language processing tasks. Inspired by this, we propose a novel dialogue generation pre-training framework to support various kinds of conversations, including chit-chat, knowledge grounded dialogues, and conversational question answering. In this framework, we adopt flexible attention mechanisms to fully leverage the bi-directional context and the uni-directional characteristic of language generation. We also introduce discrete latent variables to tackle the inherent one-to-many mapping problem in response generation. Two reciprocal tasks of response generation and latent act recognition are designed and carried out simultaneously within a shared network. Comprehensive experiments on three publicly available datasets verify the effectiveness and superiority of the proposed framework."
}
Markdown (Informal)
[PLATO: Pre-trained Dialogue Generation Model with Discrete Latent Variable](https://preview.aclanthology.org/jlcl-multiple-ingestion/2020.acl-main.9/) (Bao et al., ACL 2020)
ACL