Measuring Bias and Agreement in Large Language Model Presupposition Judgments

Katherine Atwell, Mandy Simons, Malihe Alikhani


Abstract
Identifying linguistic bias in text demands the identification not only of explicitly asserted content but also of implicit content including presuppositions. Large language models (LLMs) offer a promising automated approach to detecting presuppositions, yet the extent to which their judgments align with human intuitions remains unexplored. Moreover, LLMs may inadvertently reflect societal biases when identifying presupposed content. To empirically investigate this, we prompt multiple large language models to evaluate presuppositions across diverse textual domains, drawing from three distinct datasets annotated by human raters. We calculate the agreement between LLMs and human raters, and find several linguistic factors associated with fluctuations in human-model agreement. Our observations reveal discrepancies in human-model alignment, suggesting potential biases in LLMs, notably influenced by gender and political ideology.
Anthology ID:
2025.findings-acl.107
Volume:
Findings of the Association for Computational Linguistics: ACL 2025
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
2096–2107
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.107/
DOI:
Bibkey:
Cite (ACL):
Katherine Atwell, Mandy Simons, and Malihe Alikhani. 2025. Measuring Bias and Agreement in Large Language Model Presupposition Judgments. In Findings of the Association for Computational Linguistics: ACL 2025, pages 2096–2107, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Measuring Bias and Agreement in Large Language Model Presupposition Judgments (Atwell et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.findings-acl.107.pdf