@inproceedings{wan-etal-2022-unified,
title = "A Unified Dialogue User Simulator for Few-shot Data Augmentation",
author = "Wan, Dazhen and
Zhang, Zheng and
Zhu, Qi and
Liao, Lizi and
Huang, Minlie",
editor = "Goldberg, Yoav and
Kozareva, Zornitsa and
Zhang, Yue",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2022",
month = dec,
year = "2022",
address = "Abu Dhabi, United Arab Emirates",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2022.findings-emnlp.277/",
doi = "10.18653/v1/2022.findings-emnlp.277",
pages = "3788--3799",
abstract = "Pre-trained language models have shown superior performance in task-oriented dialogues. However, existing datasets are on limited scales, which cannot support large-scale pre-training. Fortunately, various data augmentation methods have been developed to augment large-scale task-oriented dialogue corpora. However, they heavily rely on annotated data in the target domain, which require a tremendous amount of data collection and human labeling work. In this paper, we build a unified dialogue user simulation model by pre-training on several publicly available datasets. The model can then be tuned on a target domain with few-shot data. The experiments on a target dataset across multiple domains show that our proposed model brings remarkable performance increases through data augmentation."
}
Markdown (Informal)
[A Unified Dialogue User Simulator for Few-shot Data Augmentation](https://preview.aclanthology.org/fix-sig-urls/2022.findings-emnlp.277/) (Wan et al., Findings 2022)
ACL