CREPE: Rapid Chest X-ray Report Evaluation by Predicting Multi-category Error Counts

Gihun Cho, Seunghyun Jang, Hanbin Ko, Inhyeok Baek, Chang Min Park


Abstract
We introduce CREPE (Rapid Chest X-ray Report Evaluation by Predicting Multi-category Error Counts), a rapid, interpretable, and clinically grounded metric for automated chest X-ray report generation. CREPE uses a domain-specific BERT model fine-tuned with a multi-head regression architecture to predict error counts across six clinically meaningful categories. Trained on a large-scale synthetic dataset of 32,000 annotated report pairs, CREPE demonstrates strong generalization and interpretability. On the expert-annotated ReXVal dataset, CREPE achieves a Kendall’s tau correlation of 0.786 with radiologist error counts, outperforming traditional and recent metrics. CREPE achieves these results with an inference speed approximately 280 times faster than large language model (LLM)-based approaches, enabling rapid and fine-grained evaluation for scalable development of chest X-ray report generation models.
Anthology ID:
2025.emnlp-main.1102
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
21749–21766
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1102/
DOI:
Bibkey:
Cite (ACL):
Gihun Cho, Seunghyun Jang, Hanbin Ko, Inhyeok Baek, and Chang Min Park. 2025. CREPE: Rapid Chest X-ray Report Evaluation by Predicting Multi-category Error Counts. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 21749–21766, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
CREPE: Rapid Chest X-ray Report Evaluation by Predicting Multi-category Error Counts (Cho et al., EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1102.pdf
Checklist:
 2025.emnlp-main.1102.checklist.pdf