Qinlin Zhao
2025
Exploring the Choice Behavior of Large Language Models
Weidong Wu
|
Qinlin Zhao
|
Hao Chen
|
Lexin Zhou
|
Defu Lian
|
Hong Xie
Findings of the Association for Computational Linguistics: ACL 2025
Large Language Models (LLMs) are increasingly deployed as human assistants across various domains where they help to make choices. However, the mechanisms behind LLMs’ choice behavior remain unclear, posing risks in safety-critical situations. Inspired by the intrinsic and extrinsic motivation framework within the classic human behavioral model of Self-Determination Theory and its established research methodologies, we investigate the factors influencing LLMs’ choice behavior by constructing a virtual QA platform that includes three different experimental conditions, with four models from GPT and Llama series participating in repeated experiments. Our findings indicate that LLMs’ behavior is influenced not only by intrinsic attention bias but also by extrinsic social influence, exhibiting patterns similar to the Matthew effect and Conformity. We distinguish independent pathways of these two factors in LLMs’ behavior by self-report. This work provides new insights into understanding LLMs’ behavioral patterns, exploring their human-like characteristics.
2024
AgentReview: Exploring Peer Review Dynamics with LLM Agents
Yiqiao Jin
|
Qinlin Zhao
|
Yiyang Wang
|
Hao Chen
|
Kaijie Zhu
|
Yijia Xiao
|
Jindong Wang
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Peer review is fundamental to the integrity and advancement of scientific publication. Traditional methods of peer review analyses often rely on exploration and statistics of existing peer review data, which do not adequately address the multivariate nature of the process, account for the latent variables, and are further constrained by privacy concerns due to the sensitive nature of the data. We introduce AgentReview, the first large language model (LLM) based peer review simulation framework, which effectively disentangles the impacts of multiple latent factors and addresses the privacy issue. Our study reveals significant insights, including a notable 37.1% variation in paper decisions due to reviewers’ biases, supported by sociological theories such as the social influence theory, altruism fatigue, and authority bias. We believe that this study could offer valuable insights to improve the design of peer review mechanisms.
Search
Fix author
Co-authors
- Hao Chen (陈昊) 2
- Yiqiao Jin 1
- Defu Lian 1
- Yiyang Wang 1
- Jindong Wang 1
- show all...