@inproceedings{caselli-plaza-del-arco-2025-learning,
title = "Learning from Disagreement: Entropy-Guided Few-Shot Selection for Toxic Language Detection",
author = "Caselli, Tommaso and
Plaza-del-Arco, Flor Miriam",
editor = "Calabrese, Agostina and
de Kock, Christine and
Nozza, Debora and
Plaza-del-Arco, Flor Miriam and
Talat, Zeerak and
Vargas, Francielle",
booktitle = "Proceedings of the The 9th Workshop on Online Abuse and Harms (WOAH)",
month = aug,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.woah-1.5/",
pages = "53--66",
ISBN = "979-8-89176-105-6",
abstract = "In-context learning (ICL) has shown significant benefits, particularly in scenarios where large amounts of labeled data are unavailable. However, its effectiveness for highly subjective tasks, such as toxic language detection, remains an open question. A key challenge in this setting is to select shots to maximize performance. Although previous work has focused on enhancing variety and representativeness, the role of annotator disagreement in shot selection has received less attention. In this paper, we conduct an in-depth analysis of ICL using two families of open-source LLMs (Llama-3* and Qwen2.5) of varying sizes, evaluating their performance in five prominent English datasets covering multiple toxic language phenomena. We use disaggregated annotations and categorize different types of training examples to assess their impact on model predictions. We specifically investigate whether selecting shots based on annotators' entropy {--} focusing on ambiguous or difficult examples {--} can improve generalization in LLMs. Additionally, we examine the extent to which the order of examples in prompts influences model behavior.Our results show that selecting shots based on entropy from annotator disagreement can enhance ICL performance. Specifically, ambiguous shots with a median entropy value generally lead to the best results for our selected LLMs in the few-shot setting. However, ICL often underperforms when compared to fine-tuned encoders."
}
Markdown (Informal)
[Learning from Disagreement: Entropy-Guided Few-Shot Selection for Toxic Language Detection](https://preview.aclanthology.org/landing_page/2025.woah-1.5/) (Caselli & Plaza-del-Arco, WOAH 2025)
ACL