@inproceedings{harris-etal-2024-modeling,
title = "Modeling Gender and Dialect Bias in Automatic Speech Recognition",
author = "Harris, Camille and
Mgbahurike, Chijioke and
Kumar, Neha and
Yang, Diyi",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2024.findings-emnlp.890/",
doi = "10.18653/v1/2024.findings-emnlp.890",
pages = "15166--15184",
abstract = "Dialect and gender-based biases have become an area of concern in language-dependent AI systemsincluding around automatic speech recognition (ASR) which processes speech audio into text. These potential biases raise concern for discriminatory outcomes with AI systems depending on demographic- particularly gender discrimination against women, and racial discrimination against minorities with ethnic or cultural English dialects.As such we aim to evaluate the performance of ASR systems across different genders and across dialects of English. Concretely, we take a deep dive of the performance of ASR systems on men and women across four US-based English dialects: Standard American English (SAE), African American Vernacular English (AAVE), Chicano English, and Spanglish. To do this, we construct a labeled dataset of 13 hours of podcast audio, transcribed by speakers of the represented dialects. We then evaluate zero-shot performance of different automatic speech recognition models on our dataset, and further finetune models to better understand how finetuning can impact performance. Our work fills the gap of investigating possible gender disparities within underrepresented dialects."
}
Markdown (Informal)
[Modeling Gender and Dialect Bias in Automatic Speech Recognition](https://preview.aclanthology.org/fix-sig-urls/2024.findings-emnlp.890/) (Harris et al., Findings 2024)
ACL