@inproceedings{proietti-etal-2025-machine,
title = "Has Machine Translation Evaluation Achieved Human Parity? The Human Reference and the Limits of Progress",
author = "Proietti, Lorenzo and
Perrella, Stefano and
Navigli, Roberto",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/landing_page/2025.acl-short.63/",
pages = "790--813",
ISBN = "979-8-89176-252-7",
abstract = "In Machine Translation (MT) evaluation, metric performance is assessed based on agreement with human judgments. In recent years, automatic metrics have demonstrated increasingly high levels of agreement with humans. To gain a clearer understanding of metric performance and establish an upper bound, we incorporate human baselines in the MT meta-evaluation, that is, the assessment of MT metrics' capabilities. Our results show that human annotators are not consistently superior to automatic metrics, with state-of-the-art metrics often ranking on par with or higher than human baselines. Despite these findings suggesting human parity, we discuss several reasons for caution. Finally, we explore the broader implications of our results for the research field, asking: Can we still reliably measure improvements in MT evaluation? With this work, we aim to shed light on the limits of our ability to measure progress in the field, fostering discussion on an issue that we believe is crucial to the entire MT evaluation community."
}
Markdown (Informal)
[Has Machine Translation Evaluation Achieved Human Parity? The Human Reference and the Limits of Progress](https://preview.aclanthology.org/landing_page/2025.acl-short.63/) (Proietti et al., ACL 2025)
ACL