Wanted: Personalised Bias Warnings for Gender Bias in Language Models

Chiara Di Bonaventura, Michelle Nwachukwu, Maria Stoica


Abstract
The widespread use of language models, especially Large Language Models, paired with their inherent biases can propagate and amplify societal inequalities. While research has extensively explored methods for bias mitigation and measurement, limited attention has been paid to how such biases are communicated to users, which instead can have a positive impact on increasing user trust and understanding of these models. Our study addresses this gap by investigating user preferences for gender bias mitigation, measurement and communication in language models. To this end, we conducted a user study targeting female AI practitioners with eighteen female and one male participant. Our findings reveal that user preferences for bias mitigation and measurement show strong consensus, whereas they vary widely for bias communication, underscoring the importance of tailoring warnings to individual needs.Building on these findings, we propose a framework for user-centred bias reporting, which leverages runtime monitoring techniques to assess and visualise bias in real time and in a customizable fashion.
Anthology ID:
2025.gebnlp-1.13
Volume:
Proceedings of the 6th Workshop on Gender Bias in Natural Language Processing (GeBNLP)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Agnieszka Faleńska, Christine Basta, Marta Costa-jussà, Karolina Stańczak, Debora Nozza
Venues:
GeBNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
124–136
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.gebnlp-1.13/
DOI:
Bibkey:
Cite (ACL):
Chiara Di Bonaventura, Michelle Nwachukwu, and Maria Stoica. 2025. Wanted: Personalised Bias Warnings for Gender Bias in Language Models. In Proceedings of the 6th Workshop on Gender Bias in Natural Language Processing (GeBNLP), pages 124–136, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Wanted: Personalised Bias Warnings for Gender Bias in Language Models (Di Bonaventura et al., GeBNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.gebnlp-1.13.pdf