@inproceedings{haez-dragoni-2025-neutral,
title = "Neutral Is Not Unbiased: Evaluating Implicit and Intersectional Identity Bias in {LLM}s Through Structured Narrative Scenarios",
author = "Ghanbari Haez, Saba and
Dragoni, Mauro",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.814/",
doi = "10.18653/v1/2025.findings-emnlp.814",
pages = "15060--15088",
ISBN = "979-8-89176-335-7",
abstract = "Large Language Models often reproduce societal biases, yet most evaluations overlook how such biases evolve across nuanced contexts or intersecting identities. We introduce a scenario-based evaluation framework built on 100 narrative tasks, designed to be neutral at baseline and systematically modified with gender and age cues. Grounded in the theory of Normative-Narrative Scenarios, our approach provides ethically coherent and socially plausible settings for probing model behavior. Analyzing responses from five leading LLMs{---}GPT-4o, LLaMA 3.1, Qwen2.5, Phi-4, and Mistral{---}using Critical Discourse Analysis and quantitative linguistic metrics, we find consistent evidence of bias. Gender emerges as the dominant axis of bias, with intersectional cues (e.g., age and gender combined) further intensifying disparities. Our results underscore the value of dynamic narrative progression for detecting implicit, systemic biases in Large Language Models."
}Markdown (Informal)
[Neutral Is Not Unbiased: Evaluating Implicit and Intersectional Identity Bias in LLMs Through Structured Narrative Scenarios](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.814/) (Ghanbari Haez & Dragoni, Findings 2025)
ACL