@inproceedings{qian-etal-2025-disentangling,
title = "Disentangling Biased Representations: A Causal Intervention Framework for Fairer {NLP} Models",
author = "Qian, Yangge and
Hu, Yilong and
Zhang, Siqi and
Gu, Xu and
Qin, Xiaolin",
editor = "Fale{\'n}ska, Agnieszka and
Basta, Christine and
Costa-juss{\`a}, Marta and
Sta{\'n}czak, Karolina and
Nozza, Debora",
booktitle = "Proceedings of the 6th Workshop on Gender Bias in Natural Language Processing (GeBNLP)",
month = aug,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.gebnlp-1.33/",
pages = "393--402",
ISBN = "979-8-89176-277-0",
abstract = "Natural language processing (NLP) systems often inadvertently encode and amplify social biases through entangled representations of demographic attributes and task-related attributes. To mitigate this, we propose a novel framework that combines causal analysis with practical intervention strategies. The method leverages attribute-specific prompting to isolate sensitive attributes while applying information-theoretic constraints to minimize spurious correlations. Experiments across six language models and two classification tasks demonstrate its effectiveness. We hope this work will provide the NLP community with a causal disentanglement perspective for achieving fairness in NLP systems."
}
Markdown (Informal)
[Disentangling Biased Representations: A Causal Intervention Framework for Fairer NLP Models](https://preview.aclanthology.org/landing_page/2025.gebnlp-1.33/) (Qian et al., GeBNLP 2025)
ACL