Qingqing Dong
2025
Curse of Knowledge: Your Guidance and Provided Knowledge are biasing LLM Judges in Complex Evaluation
Weiyuan Li
|
Xintao Wang
|
Siyu Yuan
|
Rui Xu
|
Jiangjie Chen
|
Qingqing Dong
|
Yanghua Xiao
|
Deqing Yang
Findings of the Association for Computational Linguistics: EMNLP 2025
As large language models (LLMs) grow more capable, they face increasingly diverse and complex tasks, making reliable evaluation challenging. The paradigm of LLMs as judges has emerged as a scalable solution, yet prior work primarily focuses on simple settings. Their reliability in complex tasks—where multi-faceted rubrics, unstructured reference answers, and nuanced criteria are critical—remains understudied. In this paper, we constructed ComplexEval Bench, a challenge benchmark designed to systematically expose and quantify Auxiliary Information Induced Biases. We systematically investigated and validated 6 previously unexplored biases across 12 basic and 3 advanced scenarios. Key findings reveal: (1) all evaluated models exhibit significant susceptibility to these biases, with bias magnitude scaling with task complexity; (2) notably, Large Reasoning Models (LRMs) show paradoxical vulnerability. Our in-depth analysis offers crucial insights for improving the accuracy and verifiability of evaluation signals, paving the way for more general and robust evaluation models.
Search
Fix author
Co-authors
- Jiangjie Chen 1
- Weiyuan Li 1
- Xintao Wang 1
- Yanghua Xiao 1
- Rui Xu 1
- show all...