@inproceedings{sousa-lopes-cardoso-2025-pam,
title = "{PAM}: Paraphrase {AMR}-Centric Evaluation Metric",
author = "Sousa, Afonso and
Lopes Cardoso, Henrique",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/display_plenaries/2025.findings-acl.879/",
pages = "17106--17121",
ISBN = "979-8-89176-256-5",
abstract = "Paraphrasing is rooted in semantics, which makes evaluating paraphrase generation systems hard. Current paraphrase generators are typically evaluated using borrowed metrics from adjacent text-to-text tasks, like machine translation or text summarization. These metrics tend to have ties to the surface form of the reference text. This is not ideal for paraphrases as we typically want variation in the lexicon while persisting semantics. To address this problem, and inspired by learned similarity evaluation on plain text, we propose PAM, a Paraphrase AMR-Centric Evaluation Metric. This metric uses AMR graphs extracted from the input text, which consist of semantic structures agnostic to the text surface form, making the resulting evaluation metric more robust to variations in syntax or lexicon. Additionally, we evaluated PAM on different semantic textual similarity datasets and found that it improves the correlations with human semantic scores when compared to other AMR-based metrics."
}
Markdown (Informal)
[PAM: Paraphrase AMR-Centric Evaluation Metric](https://preview.aclanthology.org/display_plenaries/2025.findings-acl.879/) (Sousa & Lopes Cardoso, Findings 2025)
ACL
- Afonso Sousa and Henrique Lopes Cardoso. 2025. PAM: Paraphrase AMR-Centric Evaluation Metric. In Findings of the Association for Computational Linguistics: ACL 2025, pages 17106–17121, Vienna, Austria. Association for Computational Linguistics.