Seeing Race, Feeling Bias: Emotion Stereotyping in Multimodal Language Models

Mahammed Kamruzzaman, Amanda Cercas Curry, Alba Cercas Curry, Flor Miriam Plaza-del-Arco


Abstract
Large language models (LLMs) are increasingly used to predict human emotions, but previous studies show that these models reproduce gendered emotion stereotypes. Emotion stereotypes are also tightly tied to race and skin tone (consider for example the trope of the angry black woman), but previous work has thus far overlooked this dimension. In this paper, we address this gap by introducing the first large-scale multimodal study of racial, gender, and skin-tone bias in emotion attribution, revealing how modality (text, images) and their combination shape emotion stereotypes in Multimodal LLMs (MLLMs). We evaluate four open-source MLLMs using 2.1K emotion-related events paired with 400 neutral face images across three different prompt strategies. Our findings reveal varying biases in MLLMs representations of different racial groups: models reproduce racial stereotypes across modalities, with textual cues being particularly noticeable. Models also reproduce colourist trends, with darker skin tones showing more skew. Our research highlights the need for future rigorous evaluation and mitigation strategies that account for race, colorism, and gender in MLLMs.
Anthology ID:
2025.findings-emnlp.386
Volume:
Findings of the Association for Computational Linguistics: EMNLP 2025
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
Findings
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
7317–7351
Language:
URL:
https://preview.aclanthology.org/name-variant-enfa-fane/2025.findings-emnlp.386/
DOI:
10.18653/v1/2025.findings-emnlp.386
Bibkey:
Cite (ACL):
Mahammed Kamruzzaman, Amanda Cercas Curry, Alba Cercas Curry, and Flor Miriam Plaza-del-Arco. 2025. Seeing Race, Feeling Bias: Emotion Stereotyping in Multimodal Language Models. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 7317–7351, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Seeing Race, Feeling Bias: Emotion Stereotyping in Multimodal Language Models (Kamruzzaman et al., Findings 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/name-variant-enfa-fane/2025.findings-emnlp.386.pdf
Checklist:
 2025.findings-emnlp.386.checklist.pdf