@inproceedings{supryadi-etal-2025-reprohum,
title = "{R}epro{H}um {\#}0067-01: A Reproduction of the Evaluation of Cross-Lingual Summarization",
author = "Supryadi and
Liu, Chuang and
Xiong, Deyi",
editor = "Arviv, Ofir and
Clinciu, Miruna and
Dhole, Kaustubh and
Dror, Rotem and
Gehrmann, Sebastian and
Habba, Eliya and
Itzhak, Itay and
Mille, Simon and
Perlitz, Yotam and
Santus, Enrico and
Sedoc, Jo{\~a}o and
Shmueli Scheuer, Michal and
Stanovsky, Gabriel and
Tafjord, Oyvind",
booktitle = "Proceedings of the Fourth Workshop on Generation, Evaluation and Metrics (GEM{\texttwosuperior})",
month = jul,
year = "2025",
address = "Vienna, Austria and virtual meeting",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/nschneid-patch-1/2025.gem-1.56/",
pages = "609--614",
ISBN = "979-8-89176-261-9",
abstract = "Human evaluation is crucial as it offers a nuanced understanding that automated metrics often miss. By reproducing human evaluation, we can gain a better understanding of the original results. This paper is part of the ReproHum project, where our goal is to reproduce human evaluations from previous studies. We report the reproduction results of the human evaluation of cross-lingual summarization conducted by (CITATION). By comparing the original and reproduction studies, we find that our overall evaluation findings are largely consistent with those of the previous study. However, there are notable differences in evaluation scores between the two studies for certain model outputs. These discrepancies highlight the importance of carefully selecting evaluation methodologies and human annotators."
}
Markdown (Informal)
[ReproHum #0067-01: A Reproduction of the Evaluation of Cross-Lingual Summarization](https://preview.aclanthology.org/nschneid-patch-1/2025.gem-1.56/) (Supryadi et al., GEM 2025)
ACL