@inproceedings{huang-hadfi-2025-beyond,
title = "Beyond Self-Reports: Multi-Observer Agents for Personality Assessment in Large Language Models",
author = "Huang, Yin Jou and
Hadfi, Rafik",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1150/",
doi = "10.18653/v1/2025.findings-emnlp.1150",
pages = "21086--21101",
ISBN = "979-8-89176-335-7",
abstract = "Self-report questionnaires have long been used to assess LLM personality traits, yet they fail to capture behavioral nuances due to biases and meta-knowledge contamination. This paper proposes a novel multi-observer framework for personality trait assessments in LLM agents that draws on informant-report methods in psychology. Instead of relying on self-assessments, we employ multiple observer LLM agents, each of which is configured with a specific relationship (e.g., family member, friend, or coworker). The observer agents interact with the subject LLM agent before assessing its Big Five personality traits. We show that observer-report ratings align more closely with human judgments than traditional self-reports and reveal systematic biases in LLM self-assessments. Further analysis shows that aggregating ratings of multiple observers provides more reliable results, reflecting a wisdom of the crowd effect up to 5 to 7 observers."
}Markdown (Informal)
[Beyond Self-Reports: Multi-Observer Agents for Personality Assessment in Large Language Models](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.1150/) (Huang & Hadfi, Findings 2025)
ACL