@inproceedings{arvan-parde-2024-reprohum,
title = "{R}epro{H}um {\#}0712-01: Human Evaluation Reproduction Report for {\textquotedblleft}Hierarchical Sketch Induction for Paraphrase Generation{\textquotedblright}",
author = "Arvan, Mohammad and
Parde, Natalie",
editor = "Balloccu, Simone and
Belz, Anya and
Huidrom, Rudali and
Reiter, Ehud and
Sedoc, Joao and
Thomson, Craig",
booktitle = "Proceedings of the Fourth Workshop on Human Evaluation of NLP Systems (HumEval) @ LREC-COLING 2024",
month = may,
year = "2024",
address = "Torino, Italia",
publisher = "ELRA and ICCL",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.humeval-1.18/",
pages = "210--220",
abstract = "Human evaluations are indispensable in the development of NLP systems because they provide direct insights into how effectively these systems meet real-world needs and expectations. Ensuring the reproducibility of these evaluations is vital for maintaining credibility in natural language processing research. This paper presents our reproduction of the human evaluation experiments conducted by Hosking et al. (2022) for their paraphrase generation approach. Through careful replication we found that our results closely align with those in the original study, indicating a high degree of reproducibility."
}
Markdown (Informal)
[ReproHum #0712-01: Human Evaluation Reproduction Report for “Hierarchical Sketch Induction for Paraphrase Generation”](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.humeval-1.18/) (Arvan & Parde, HumEval 2024)
ACL