2025
pdf
bib
abs
From Generation to Judgment: Opportunities and Challenges of LLM-as-a-judge
Dawei Li
|
Bohan Jiang
|
Liangjie Huang
|
Alimohammad Beigi
|
Chengshuai Zhao
|
Zhen Tan
|
Amrita Bhattacharjee
|
Yuxuan Jiang
|
Canyu Chen
|
Tianhao Wu
|
Kai Shu
|
Lu Cheng
|
Huan Liu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Assessment and evaluation have long been critical challenges in artificial intelligence (AI) and natural language processing (NLP). Traditional methods, usually matching-based or small model-based, often fall short in open-ended and dynamic scenarios. Recent advancements in Large Language Models (LLMs) inspire the “LLM-as-a-judge” paradigm, where LLMs are leveraged to perform scoring, ranking, or selection for various machine learning evaluation scenarios. This paper presents a comprehensive survey of LLM-based judgment and assessment, offering an in-depth overview to review this evolving field. We first provide the definition from both input and output perspectives. Then we introduce a systematic taxonomy to explore LLM-as-a-judge along three dimensions: what to judge, how to judge, and how to benchmark. Finally, we also highlight key challenges and promising future directions for this emerging area.
pdf
bib
abs
A Simple Yet Effective Method for Non-Refusing Context Relevant Fine-grained Safety Steering in LLMs
Shaona Ghosh
|
Amrita Bhattacharjee
|
Yftah Ziser
|
Christopher Parisien
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Fine-tuning large language models (LLMs) to meet evolving safety policies is costly and impractical. Mechanistic interpretability enables inference-time control through latent activation steering, but its potential for precise, customizable safety adjustments remains underexplored. We propose SafeSteer, a simple and effective method to guide LLM outputs by (i) leveraging category-specific steering vectors for fine-grained control, (ii) applying a gradient-free, unsupervised approach that enhances safety while preserving text quality and topic relevance without forcing explicit refusals, and (iii) eliminating the need for contrastive safe data. Across multiple LLMs, datasets, and risk categories, SafeSteer provides precise control, avoids blanket refusals, and directs models to generate safe, relevant content, aligning with recent findings that simple activation-steering techniques often outperform more complex alternatives.
2024
pdf
bib
abs
Large Language Models for Data Annotation and Synthesis: A Survey
Zhen Tan
|
Dawei Li
|
Song Wang
|
Alimohammad Beigi
|
Bohan Jiang
|
Amrita Bhattacharjee
|
Mansooreh Karami
|
Jundong Li
|
Lu Cheng
|
Huan Liu
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Data annotation and synthesis generally refers to the labeling or generating of raw data with relevant information, which could be used for improving the efficacy of machine learning models. The process, however, is labor-intensive and costly. The emergence of advanced Large Language Models (LLMs), exemplified by GPT-4, presents an unprecedented opportunity to automate the complicated process of data annotation and synthesis. While existing surveys have extensively covered LLM architecture, training, and general applications, we uniquely focus on their specific utility for data annotation. This survey contributes to three core aspects: LLM-Based Annotation Generation, LLM-Generated Annotations Assessment, and LLM-Generated Annotations Utilization. Furthermore, this survey includes an in-depth taxonomy of data types that LLMs can annotate, a comprehensive review of learning strategies for models utilizing LLM-generated annotations, and a detailed discussion of the primary challenges and limitations associated with using LLMs for data annotation and synthesis. Serving as a key guide, this survey aims to assist researchers and practitioners in exploring the potential of the latest LLMs for data annotation, thereby fostering future advancements in this critical field.
pdf
bib
Defending Against Social Engineering Attacks in the Age of LLMs
Lin Ai
|
Tharindu Sandaruwan Kumarage
|
Amrita Bhattacharjee
|
Zizhou Liu
|
Zheng Hui
|
Michael S. Davinroy
|
James Cook
|
Laura Cassani
|
Kirill Trapeznikov
|
Matthias Kirchner
|
Arslan Basharat
|
Anthony Hoogs
|
Joshua Garland
|
Huan Liu
|
Julia Hirschberg
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
pdf
bib
abs
Towards Interpretable Hate Speech Detection using Large Language Model-extracted Rationales
Ayushi Nirmal
|
Amrita Bhattacharjee
|
Paras Sheth
|
Huan Liu
Proceedings of the 8th Workshop on Online Abuse and Harms (WOAH 2024)
Although social media platforms are a prominent arena for users to engage in interpersonal discussions and express opinions, the facade and anonymity offered by social media may allow users to spew hate speech and offensive content. Given the massive scale of such platforms, there arises a need to automatically identify and flag instances of hate speech. Although several hate speech detection methods exist, most of these black-box methods are not interpretable or explainable by design. To address the lack of interpretability, in this paper, we propose to use state-of-the-art Large Language Models (LLMs) to extract features in the form of rationales from the input text, to train a base hate speech classifier, thereby enabling faithful interpretability by design. Our framework effectively combines the textual understanding capabilities of LLMs and the discriminative power of state-of-the-art hate speech classifiers to make these classifiers faithfully interpretable. Our comprehensive evaluation on a variety of social media hate speech datasets demonstrate: (1) the goodness of the LLM-extracted rationales, and (2) the surprising retention of detector performance even after training to ensure interpretability. All code and data will be made available at https://github.com/AmritaBh/shield.
2023
pdf
bib
J-Guard: Journalism Guided Adversarially Robust Detection of AI-generated News
Tharindu Kumarage
|
Amrita Bhattacharjee
|
Djordje Padejski
|
Kristy Roschke
|
Dan Gillmor
|
Scott Ruston
|
Huan Liu
|
Joshua Garland
Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
pdf
bib
ConDA: Contrastive Domain Adaptation for AI-generated Text Detection
Amrita Bhattacharjee
|
Tharindu Kumarage
|
Raha Moraffah
|
Huan Liu
Proceedings of the 13th International Joint Conference on Natural Language Processing and the 3rd Conference of the Asia-Pacific Chapter of the Association for Computational Linguistics (Volume 1: Long Papers)
pdf
bib
abs
Towards Detecting Harmful Agendas in News Articles
Melanie Subbiah
|
Amrita Bhattacharjee
|
Yilun Hua
|
Tharindu Kumarage
|
Huan Liu
|
Kathleen McKeown
Proceedings of the 13th Workshop on Computational Approaches to Subjectivity, Sentiment, & Social Media Analysis
Manipulated news online is a growing problem which necessitates the use of automated systems to curtail its spread. We argue that while misinformation and disinformation detection have been studied, there has been a lack of investment in the important open challenge of detecting harmful agendas in news articles; identifying harmful agendas is critical to flag news campaigns with the greatest potential for real world harm. Moreover, due to real concerns around censorship, harmful agenda detectors must be interpretable to be effective. In this work, we propose this new task and release a dataset, NewsAgendas, of annotated news articles for agenda identification. We show how interpretable systems can be effective on this task and demonstrate that they can perform comparably to black-box models.