Eshwar Chandrasekharan
2025
SLM-Mod: Small Language Models Surpass LLMs at Content Moderation
Xianyang Zhan
|
Agam Goyal
|
Yilun Chen
|
Eshwar Chandrasekharan
|
Koustuv Saha
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Large language models (LLMs) have shown promise in many natural language understanding tasks, including content moderation. However, these models can be expensive to query in real-time and do not allow for a community-specific approach to content moderation. To address these challenges, we explore the use of open-source small language models (SLMs) for community-specific content moderation tasks. We fine-tune and evaluate SLMs (less than 15B parameters) by comparing their performance against much larger open- and closed-sourced models in both a zero-shot and few-shot setting. Using 150K comments from 15 popular Reddit communities, we find that SLMs outperform zero-shot LLMs at content moderation-11.5% higher accuracy and 25.7% higher recall on average across all communities. Moreover, few-shot in-context learning leads to only a marginal increase in the performance of LLMs, still lacking compared to SLMs. We further show the promise of cross-community content moderation, which has implications for new communities and the development of cross-platform moderation techniques. Finally, we outline directions for future work on language model based content moderation.
2019
A Just and Comprehensive Strategy for Using NLP to Address Online Abuse
David Jurgens
|
Libby Hemphill
|
Eshwar Chandrasekharan
Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics
Online abusive behavior affects millions and the NLP community has attempted to mitigate this problem by developing technologies to detect abuse. However, current methods have largely focused on a narrow definition of abuse to detriment of victims who seek both validation and solutions. In this position paper, we argue that the community needs to make three substantive changes: (1) expanding our scope of problems to tackle both more subtle and more serious forms of abuse, (2) developing proactive technologies that counter or inhibit abuse before it harms, and (3) reframing our effort within a framework of justice to promote healthy communities.
Search
Fix data
Co-authors
- Yilun Chen 1
- Agam Goyal 1
- Libby Hemphill 1
- David Jurgens 1
- Koustuv Saha 1
- show all...