@inproceedings{soubki-rambow-2025-machine,
title = "Machine Theory of Mind Needs Machine Validation",
author = "Soubki, Adil and
Rambow, Owen",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.951/",
doi = "10.18653/v1/2025.findings-acl.951",
pages = "18495--18505",
ISBN = "979-8-89176-256-5",
abstract = "In the last couple years, there has been a flood of interest in studying the extent to which language models (LMs) have a theory of mind (ToM) {---} the ability to ascribe mental states to themselves and others. The results provide an unclear picture of the current state of the art, with some finding near-human performance and others near-zero. To make sense of this landscape, we perform a survey of 16 recent studies aimed at measuring ToM in LMs and find that, while almost all perform checks for human identifiable issues, less than half do so for patterns only a machine might exploit. Among those that do perform such validation, which we call machine validation, none identify LMs to exceed human performance. We conclude that the datasets that show high LM performance on ToM tasks are easier than their peers, likely due to the presence of spurious patterns in the data, and we caution against building ToM benchmarks relying solely on human validation of the data."
}
Markdown (Informal)
[Machine Theory of Mind Needs Machine Validation](https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.951/) (Soubki & Rambow, Findings 2025)
ACL