Jinglu Hu
2025
MaXIFE: Multilingual and Cross-lingual Instruction Following Evaluation
Yile Liu
|
Ziwei Ma
|
Xiu Jiang
|
Jinglu Hu
|
ChangJing ChangJing
|
Liang Li
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
With the rapid adoption of large language models (LLMs) in natural language processing, the ability to follow instructions has emerged as a key metric for evaluating their practical utility. However, existing evaluation methods often focus on single-language scenarios, overlooking the challenges and differences present in multilingual and cross-lingual contexts. To address this gap, we introduce MaXIFE: a comprehensive evaluation benchmark designed to assess instruction-following capabilities across 23 different languages with 1667 verifiable instruction tasks. MaXIFE integrates both Rule-Based Evaluation and Model-Based Evaluation, ensuring a balance of efficiency and accuracy. We applied MaXIFE to evaluate several leading commercial LLMs, establishing baseline results for future comparisons. By providing a standardized tool for multilingual instruction-following evaluation, MaXIFE aims to advance research and development in natural language processing.
2020
Improving Image Captioning Evaluation by Considering Inter References Variance
Yanzhi Yi
|
Hangyu Deng
|
Jinglu Hu
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
Evaluating image captions is very challenging partially due to the fact that there are multiple correct captions for every single image. Most of the existing one-to-one metrics operate by penalizing mismatches between reference and generative caption without considering the intrinsic variance between ground truth captions. It usually leads to over-penalization and thus a bad correlation to human judgment. Recently, the latest one-to-one metric BERTScore can achieve high human correlation in system-level tasks while some issues can be fixed for better performance. In this paper, we propose a novel metric based on BERTScore that could handle such a challenge and extend BERTScore with a few new features appropriately for image captioning evaluation. The experimental results show that our metric achieves state-of-the-art human judgment correlation.
Search
Fix author
Co-authors
- ChangJing ChangJing 1
- Hangyu Deng 1
- Xiu Jiang 1
- Liang Li 1
- Yile Liu 1
- show all...
Venues
- acl2