@inproceedings{atwell-etal-2025-measuring,
title = "Measuring Bias and Agreement in Large Language Model Presupposition Judgments",
author = "Atwell, Katherine and
Simons, Mandy and
Alikhani, Malihe",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.107/",
pages = "2096--2107",
ISBN = "979-8-89176-256-5",
abstract = "Identifying linguistic bias in text demands the identification not only of explicitly asserted content but also of implicit content including presuppositions. Large language models (LLMs) offer a promising automated approach to detecting presuppositions, yet the extent to which their judgments align with human intuitions remains unexplored. Moreover, LLMs may inadvertently reflect societal biases when identifying presupposed content. To empirically investigate this, we prompt multiple large language models to evaluate presuppositions across diverse textual domains, drawing from three distinct datasets annotated by human raters. We calculate the agreement between LLMs and human raters, and find several linguistic factors associated with fluctuations in human-model agreement. Our observations reveal discrepancies in human-model alignment, suggesting potential biases in LLMs, notably influenced by gender and political ideology."
}
Markdown (Informal)
[Measuring Bias and Agreement in Large Language Model Presupposition Judgments](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.107/) (Atwell et al., Findings 2025)
ACL