Evaluating the Effect of Retrieval Augmentation on Social Biases

Tianhui Zhang, Yi Zhou, Danushka Bollegala


Abstract
Retrieval Augmented Generation (RAG) is a popular method for injecting up-to-date information into Large Language Model (LLM)-based Natural Language Generation (NLG) systems. While RAG can enhance factual accuracy, its effect on the social biases inherent in LLMs is not well understood. This paper systematically investigates how RAG modulates social biases across three languages (English, Japanese, and Chinese) and four categories (gender, race, age, and religion). By evaluating various generator LLMs on the BBQ benchmark, we analyse how document collections with controlled stereotypical content affect RAG outputs. We find that biases present in the retrieved documents are often significantly amplified in the generated texts, even when the base LLM itself has a low-level of intrinsic bias. These findings raise concerns about the social fairness of RAG systems, underscoring the urgent need for careful bias evaluation before real-world deployment.
Anthology ID:
2026.eacl-long.233
Volume:
Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
March
Year:
2026
Address:
Rabat, Morocco
Editors:
Vera Demberg, Kentaro Inui, Lluís Marquez
Venue:
EACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
5004–5026
Language:
URL:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.233/
DOI:
Bibkey:
Cite (ACL):
Tianhui Zhang, Yi Zhou, and Danushka Bollegala. 2026. Evaluating the Effect of Retrieval Augmentation on Social Biases. In Proceedings of the 19th Conference of the European Chapter of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5004–5026, Rabat, Morocco. Association for Computational Linguistics.
Cite (Informal):
Evaluating the Effect of Retrieval Augmentation on Social Biases (Zhang et al., EACL 2026)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-eacl/2026.eacl-long.233.pdf