@inproceedings{chen-etal-2023-natural,
title = "Natural Response Generation for {C}hinese Reading Comprehension",
author = "Chen, Nuo and
Li, Hongguang and
Bao, Yinan and
Wang, Baoyuan and
Li, Jia",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-emnlp.739/",
doi = "10.18653/v1/2023.findings-emnlp.739",
pages = "11068--11081",
abstract = "Machine reading comprehension (MRC) is an important area of conversation agents and draws a lot of attention. However, there is a notable limitation to current MRC benchmarks: The labeled answers are mostly either spans extracted from the target corpus or the choices of the given candidates, ignoring the natural aspect of high-quality responses. As a result, MRC models trained on these datasets can not generate human-like responses in real QA scenarios. To this end, we construct a new dataset called \textbf{Penguin} to promote the research of MRC, providing a training and test bed for natural response generation to real scenarios. Concretely, Penguin consists of 200k training data with high-quality fluent, and well-informed responses. Penguin is the first benchmark towards natural response generation in Chinese MRC on a relatively large scale. To address the challenges in Penguin, we develop two strong baselines: end-to-end and two-stage frameworks. Following that, we further design \textit{Prompt-BART}: fine-tuning the pre-trained generative language models with a mixture of prefix prompts in Penguin. Extensive experiments validated the effectiveness of this design."
}
Markdown (Informal)
[Natural Response Generation for Chinese Reading Comprehension](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-emnlp.739/) (Chen et al., Findings 2023)
ACL