@inproceedings{cheng-amiri-2024-fairflow,
title = "{F}air{F}low: Mitigating Dataset Biases through Undecided Learning for Natural Language Understanding",
author = "Cheng, Jiali and
Amiri, Hadi",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.1225/",
doi = "10.18653/v1/2024.emnlp-main.1225",
pages = "21960--21975",
abstract = "Language models are prone to dataset biases, known as shortcuts and spurious correlations in data, which often result in performance drop on new data. We present a new debiasing framework called FairFlow that mitigates dataset biases by learning to be \textit{undecided} in its predictions for data samples or representations associated with known or unknown biases. The framework introduces two key components: a suite of data and model perturbation operations that generate different biased views of input samples, and a contrastive objective that learns debiased and robust representations from the resulting biased views of samples. Experiments show that FairFlow outperforms existing debiasing methods, particularly against out-of-domain and hard test samples without compromising the in-domain performance."
}
Markdown (Informal)
[FairFlow: Mitigating Dataset Biases through Undecided Learning for Natural Language Understanding](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.emnlp-main.1225/) (Cheng & Amiri, EMNLP 2024)
ACL