@inproceedings{prakash-lee-2023-layered,
title = "Layered Bias: Interpreting Bias in Pretrained Large Language Models",
author = "Prakash, Nirmalendu and
Lee, Roy Ka-Wei",
editor = "Belinkov, Yonatan and
Hao, Sophie and
Jumelet, Jaap and
Kim, Najoung and
McCarthy, Arya and
Mohebbi, Hosein",
booktitle = "Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2023.blackboxnlp-1.22/",
doi = "10.18653/v1/2023.blackboxnlp-1.22",
pages = "284--295",
abstract = "Large language models (LLMs) like GPT and PALM have excelled in numerous natural language processing (NLP) tasks such as text generation, question answering, and translation. However, they are also found to have inherent social biases. To address this, recent studies have proposed debiasing techniques like iterative nullspace projection (INLP) and Counterfactual Data Augmentation (CDA). Additionally, there{'}s growing interest in understanding the intricacies of these models. Some researchers focus on individual neural units, while others examine specific layers. In our study, we benchmark newly released models, assess the impact of debiasing methods, and investigate how biases are linked to different transformer layers using a method called Logit Lens. Specifically, we evaluate three modern LLMs: OPT, LLaMA, and LLaMA2, and their debiased versions. Our experiments are based on two popular bias evaluation datasets, StereoSet and CrowS-Pairs, and we perform a layer-by-layer analysis using the Logit Lens."
}
Markdown (Informal)
[Layered Bias: Interpreting Bias in Pretrained Large Language Models](https://preview.aclanthology.org/fix-sig-urls/2023.blackboxnlp-1.22/) (Prakash & Lee, BlackboxNLP 2023)
ACL