Machine Theory of Mind Needs Machine Validation

Adil Soubki, Owen Rambow


Abstract
In the last couple years, there has been a flood of interest in studying the extent to which language models (LMs) have a theory of mind (ToM) — the ability to ascribe mental states to themselves and others. The results provide an unclear picture of the current state of the art, with some finding near-human performance and others near-zero. To make sense of this landscape, we perform a survey of 16 recent studies aimed at measuring ToM in LMs and find that, while almost all perform checks for human identifiable issues, less than half do so for patterns only a machine might exploit. Among those that do perform such validation, which we call machine validation, none identify LMs to exceed human performance. We conclude that the datasets that show high LM performance on ToM tasks are easier than their peers, likely due to the presence of spurious patterns in the data, and we caution against building ToM benchmarks relying solely on human validation of the data.
Anthology ID:
2025.findings-acl.951
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
18495–18505
Language:
URL:
https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.951/
DOI:
10.18653/v1/2025.findings-acl.951
Bibkey:
Cite (ACL):
Adil Soubki and Owen Rambow. 2025. Machine Theory of Mind Needs Machine Validation. In Findings of the Association for Computational Linguistics: ACL 2025, pages 18495–18505, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Machine Theory of Mind Needs Machine Validation (Soubki & Rambow, Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/mtsummit-25-ingestion/2025.findings-acl.951.pdf