@inproceedings{yu-etal-2025-analysis,
title = "An Analysis of Large Language Models for Simulating User Responses in Surveys",
author = "Yu, Ziyun and
Zhou, Yiru and
Zhao, Chen and
Wen, Hongyi",
editor = "Inui, Kentaro and
Sakti, Sakriani and
Wang, Haofen and
Wong, Derek F. and
Bhattacharyya, Pushpak and
Banerjee, Biplab and
Ekbal, Asif and
Chakraborty, Tanmoy and
Singh, Dhirendra Pratap",
booktitle = "Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics",
month = dec,
year = "2025",
address = "Mumbai, India",
publisher = "The Asian Federation of Natural Language Processing and The Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.ijcnlp-short.22/",
pages = "242--259",
ISBN = "979-8-89176-299-2",
abstract = "Using Large Language Models (LLMs) to simulate user opinions has received growing attention. Yet LLMs, especially trained with reinforcement learning from human feedback (RLHF), are known to exhibit biases toward dominant viewpoints, raising concerns about their ability to represent users from diverse demographic and cultural backgrounds. In this work, we examine the extent to which LLMs can simulate human responses to cross-domain survey questions and propose two LLM-based approaches: chain-of-thought (COT) prompting and Diverse Claims Generation (CLAIMSIM), which elicits viewpoints from LLM parametric knowledge as contextual input. Experiments on the survey question answering task indicate that, while CLAIMSIM produces more diverse responses, both approaches struggle to accurately simulate users. Further analysis reveals two key limitations: (1) LLMs tend to maintain fixed viewpoints across varying demographic features, and generate single-perspective claims; and (2) when presented with conflicting claims, LLMs struggle to reason over nuanced differences among demographic features, limiting their ability to adapt responses to specific user profiles."
}Markdown (Informal)
[An Analysis of Large Language Models for Simulating User Responses in Surveys](https://preview.aclanthology.org/ingest-ijcnlp-aacl/2025.ijcnlp-short.22/) (Yu et al., IJCNLP-AACL 2025)
ACL
- Ziyun Yu, Yiru Zhou, Chen Zhao, and Hongyi Wen. 2025. An Analysis of Large Language Models for Simulating User Responses in Surveys. In Proceedings of the 14th International Joint Conference on Natural Language Processing and the 4th Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics, pages 242–259, Mumbai, India. The Asian Federation of Natural Language Processing and The Association for Computational Linguistics.