@inproceedings{chen-etal-2023-automatic,
title = "Automatic Evaluate Dialogue Appropriateness by Using Dialogue Act",
author = "Chen, Bao and
Wang, Yuanjie and
Liu, Zeming and
Guo, Yuhang",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.findings-emnlp.492/",
doi = "10.18653/v1/2023.findings-emnlp.492",
pages = "7361--7372",
abstract = "Evaluation of dialogue systems requires assessing various aspects, among which appropriateness holds significance as a core element of communicative language competence. However, current evaluations heavily rely on human judgments, which are time-consuming, labor-intensive, prone to biases, and lacking objectivity. In this paper, we introduce Dialogue Act Appropriateness (DAA), a novel method that utilizes the underlying patterns of dialogue act transitions to evaluate the appropriateness of chatbot responses. We learn transition patterns from human-human dialogue corpora, evaluating chatbot appropriateness by measuring the similarity of their transition patterns to those observed in human-human dialogues. To validate DAA, we annotate a test dataset by manually evaluating the appropriateness of dialogues from multiple chatbot systems. The experimental results demonstrate a strong correlation between our evaluation metric and human ratings, establishing the reliability of DAA as a measure of dialogue appropriateness."
}
Markdown (Informal)
[Automatic Evaluate Dialogue Appropriateness by Using Dialogue Act](https://preview.aclanthology.org/fix-sig-urls/2023.findings-emnlp.492/) (Chen et al., Findings 2023)
ACL