@inproceedings{fu-liu-2025-reliable,
title = "How Reliable is Multilingual {LLM}-as-a-Judge?",
author = "Fu, Xiyan and
Liu, Wei",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2025",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.587/",
doi = "10.18653/v1/2025.findings-emnlp.587",
pages = "11040--11053",
ISBN = "979-8-89176-335-7",
abstract = "LLM-as-a-Judge has emerged as a popular evaluation strategy, where advanced large language models assess generation results in alignment with human instructions. While these models serve as a promising alternative to human annotators, their reliability in multilingual evaluation remains uncertain. To bridge this gap, we conduct a comprehensive analysis of multilingual LLM-as-a-Judge. Specifically, we evaluate five models from different model families across five diverse tasks involving 25 languages. Our findings reveal that LLMs struggle to achieve consistent judgment results across languages, with an average Fleiss' Kappa of approximately 0.3, and some models performing even worse. To investigate the cause of inconsistency, we analyze various influencing factors. We observe that consistency varies significantly across languages, with particularly poor performance in low-resource languages. Additionally, we find that neither training on multilingual data nor increasing model scale directly improves judgment consistency. These findings suggest that LLMs are not yet reliable for evaluating multilingual predictions. Our work provides valuable insights into the limitations of multilingual LLM-as-a-Judge, and sheds light on future research."
}Markdown (Informal)
[How Reliable is Multilingual LLM-as-a-Judge?](https://preview.aclanthology.org/author-page-yu-wang-polytechnic/2025.findings-emnlp.587/) (Fu & Liu, Findings 2025)
ACL
- Xiyan Fu and Wei Liu. 2025. How Reliable is Multilingual LLM-as-a-Judge?. In Findings of the Association for Computational Linguistics: EMNLP 2025, pages 11040–11053, Suzhou, China. Association for Computational Linguistics.