ArGAN: Arabic Gender, Ability, and Nationality Dataset for Evaluating Biases in Large Language Models

Ranwa Aly, Yara Allam, Rana Gaber, Christine Basta


Abstract
Large language models (LLMs) are pretrained on substantial, unfiltered corpora, assembled from a variety of sources. This risks inheriting the deep-rooted biases that exist within them, both implicit and explicit. This is even more apparent in low-resource languages, where corpora may be prioritized by quantity over quality, potentially leading to more unchecked biases. More specifically, we address the biases present in the Arabic language in both general-purpose and Arabic-specialized architectures in three dimensions of demographics: gender, ability, and nationality. To properly assess the fairness of these models, we experiment with bias-revealing prompts and estimate the performance using existing evaluation metrics, and propose adaptations to others.
Anthology ID:
2025.gebnlp-1.23
Volume:
Proceedings of the 6th Workshop on Gender Bias in Natural Language Processing (GeBNLP)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Agnieszka Faleńska, Christine Basta, Marta Costa-jussà, Karolina Stańczak, Debora Nozza
Venues:
GeBNLP | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
256–267
Language:
URL:
https://preview.aclanthology.org/display_plenaries/2025.gebnlp-1.23/
DOI:
Bibkey:
Cite (ACL):
Ranwa Aly, Yara Allam, Rana Gaber, and Christine Basta. 2025. ArGAN: Arabic Gender, Ability, and Nationality Dataset for Evaluating Biases in Large Language Models. In Proceedings of the 6th Workshop on Gender Bias in Natural Language Processing (GeBNLP), pages 256–267, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
ArGAN: Arabic Gender, Ability, and Nationality Dataset for Evaluating Biases in Large Language Models (Aly et al., GeBNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/display_plenaries/2025.gebnlp-1.23.pdf