@inproceedings{cachola-etal-2025-evaluating,
title = "Evaluating the Evaluators: Are readability metrics good measures of readability?",
author = "Cachola, Isabel and
Khashabi, Daniel and
Dredze, Mark",
editor = "Christodoulopoulos, Christos and
Chakraborty, Tanmoy and
Rose, Carolyn and
Peng, Violet",
booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2025",
address = "Suzhou, China",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1225/",
pages = "24022--24038",
ISBN = "979-8-89176-332-6",
abstract = "Plain Language Summarization (PLS) aims to distill complex documents into accessible summaries for non-expert audiences. In this paper, we conduct a thorough survey of PLS literature, and identify that the current standard practice for readability evaluation is to use traditional readability metrics, such as Flesch-Kincaid Grade Level (FKGL). However, despite proven utility in other fields, these metrics have not been compared to human readability judgments in PLS. We evaluate 8 readability metrics and show that most correlate poorly with human judgments, including the most popular metric, FKGL. We then show that Language Models (LMs) are better judges of readability, with the best-performing model achieving a Pearson correlation of 0.56 with human judgments. Extending our analysis to PLS datasets, which contain summaries aimed at non-expert audiences, we find that LMs better capture deeper measures of readability, such as required background knowledge, and lead to different conclusions than the traditional metrics. Based on these findings, we offer recommendations for best practices in the evaluation of plain language summaries."
}Markdown (Informal)
[Evaluating the Evaluators: Are readability metrics good measures of readability?](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1225/) (Cachola et al., EMNLP 2025)
ACL