Samuel Bell
2025
On the Role of Speech Data in Reducing Toxicity Detection Bias
Samuel Bell
|
Mariano Coria Meglioli
|
Megan Richards
|
Eduardo Sánchez
|
Christophe Ropers
|
Skyler Wang
|
Adina Williams
|
Levent Sagun
|
Marta R. Costa-jussà
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Text toxicity detection systems exhibit significant biases, producing disproportionate rates of false positives on samples mentioning demographic groups. But what about toxicity detection in speech? To investigate the extent to which text-based biases are mitigated by speech-based systems, we produce a set of high-quality group annotations for the multilingual MuTOX dataset, and then leverage these annotations to systematically compare speech- and text-based toxicity classifiers. Our findings indicate that access to speech data during inference supports reduced bias against group mentions, particularly for ambiguous and disagreement-inducing samples. Our results also suggest that improving classifiers, rather than transcription pipelines, is more helpful for reducing group bias. We publicly release our annotations and provide recommendations for future toxicity dataset construction.
2019
Context is Key: Grammatical Error Detection with Contextual Word Representations
Samuel Bell
|
Helen Yannakoudakis
|
Marek Rei
Proceedings of the Fourteenth Workshop on Innovative Use of NLP for Building Educational Applications
Grammatical error detection (GED) in non-native writing requires systems to identify a wide range of errors in text written by language learners. Error detection as a purely supervised task can be challenging, as GED datasets are limited in size and the label distributions are highly imbalanced. Contextualized word representations offer a possible solution, as they can efficiently capture compositional information in language and can be optimized on large amounts of unsupervised data. In this paper, we perform a systematic comparison of ELMo, BERT and Flair embeddings (Peters et al., 2017; Devlin et al., 2018; Akbik et al., 2018) on a range of public GED datasets, and propose an approach to effectively integrate such representations in current methods, achieving a new state of the art on GED. We further analyze the strengths and weaknesses of different contextual embeddings for the task at hand, and present detailed analyses of their impact on different types of errors.
Search
Fix data
Co-authors
- Marta R. Costa-jussà 1
- Mariano Coria Meglioli 1
- Marek Rei 1
- Megan Richards 1
- Christophe Ropers 1
- show all...