@inproceedings{wu-etal-2025-x,
title = "{X}-{TURING}: Towards an Enhanced and Efficient {T}uring Test for Long-Term Dialogue Agents",
author = "Wu, Weiqi and
Wu, Hongqiu and
Zhao, Hai",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.293/",
pages = "5874--5889",
ISBN = "979-8-89176-251-0",
abstract = "The Turing test examines whether AIs exhibit human-like behaviour in natural language conversations. The traditional setting limits each participant to one message at a time and requires constant human participation. This fails to reflect a natural conversational style and hinders the evaluation of dialogue agents based on Large Language Models (LLMs) in complex and prolonged interactions. This paper proposes X-Turing, which enhances the original test with a burst dialogue pattern, allowing more dynamic exchanges using consecutive messages. It further reduces human workload by iteratively generating dialogues that simulate the long-term interaction between the agent and a human to compose the majority of the test process. With the pseudo-dialogue history, the agent then engages in a shorter dialogue with a real human, which is paired with a human-human conversation on the same topic to be judged using questionnaires. We introduce the X-Turn Pass-Rate metric to assess the human likeness of LLMs across varying durations. While LLMs like GPT-4 initially perform well, achieving pass rates of 51.9{\%} and 38.9{\%} during 3 turns and 10 turns of dialogues respectively, their performance drops as the dialogue progresses, which underscores the difficulty in maintaining consistency in the long term."
}
Markdown (Informal)
[X-TURING: Towards an Enhanced and Efficient Turing Test for Long-Term Dialogue Agents](https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.293/) (Wu et al., ACL 2025)
ACL