Martin Semmann
2025
HatePRISM: Policies, Platforms, and Research Integration. Advancing NLP for Hate Speech Proactive Mitigation
Naquee Rizwan
|
Seid Muhie Yimam
|
Daryna Dementieva
|
Dr. Florian Skupin
|
Tim Fischer
|
Daniil Moskovskiy
|
Aarushi Ajay Borkar
|
Robert Geislinger
|
Punyajoy Saha
|
Sarthak Roy
|
Martin Semmann
|
Alexander Panchenko
|
Chris Biemann
|
Animesh Mukherjee
Findings of the Association for Computational Linguistics: ACL 2025
Despite regulations imposed by nations and social media platforms, e.g. (Government of India, 2021; European Parliament and Council of the European Union, 2022), inter alia, hateful content persists as a significant challenge. Existing approaches primarily rely on reactive measures such as blocking or suspending offensive messages, with emerging strategies focusing on proactive measurements like detoxification and counterspeech. In our work, which we call HATEPRISM, we conduct a comprehensive examination of hate speech regulations and strategies from three perspectives: country regulations, social platform policies, and NLP research datasets. Our findings reveal significant inconsistencies in hate speech definitions and moderation practices across jurisdictions and platforms, alongside a lack of alignment with research efforts. Based on these insights, we suggest ideas and research direction for further exploration of a unified framework for automated hate speech moderation incorporating diverse strategies.
2024
UHH at AVeriTeC: RAG for Fact-Checking with Real-World Claims
Özge Sevgili
|
Irina Nikishina
|
Seid Muhie Yimam
|
Martin Semmann
|
Chris Biemann
Proceedings of the Seventh Fact Extraction and VERification Workshop (FEVER)
This paper presents UHH’s approach developed for the AVeriTeC shared task. The goal of the challenge is to verify given real-world claims with evidences from the Web. In this shared task, we investigate a Retrieval-Augmented Generation (RAG) model, which mainly contains retrieval, generation, and augmentation components. We start with the selection of the top 10k evidences via BM25 scores, and continue with two approaches to retrieve the most similar evidences: (1) to retrieve top 10 evidences through vector similarity, generate questions for them, and rerank them or (2) to generate questions for the claim and retrieve the most similar evidence, again, through vector similarity. After retrieving the top evidences, a Large Language Model (LLM) is prompted using the claim along with either all evidences or individual evidence to predict the label. Our system submission, UHH, using the first approach and individual evidence prompts, ranks 6th out of 23 systems.