EditInspector: A Benchmark for Evaluation of Text-Guided Image Edits

Ron Yosef, Yonatan Bitton, Dani Lischinski, Moran Yanuka


Abstract
Text-guided image editing, fueled by recent advancements in generative AI, is becoming increasingly widespread. This trend highlights the need for a comprehensive framework to verify text-guided edits and assess their quality. To address this need, we introduce EditInspector, a novel benchmark for evaluation of text-guided image edits, based on human annotations collected using an extensive template for edit verification. We leverage EditInspector to evaluate the performance of state-of-the-art (SoTA) vision and language models in assessing edits across various dimensions, including accuracy, artifact detection, visual quality, seamless integration with the image scene, adherence to common sense, and the ability to describe edit-induced changes. Our findings indicate that current models struggle to evaluate edits comprehensively and frequently hallucinate when describing the changes. To address these challenges, we propose two novel methods that outperform SoTA models in both artifact detection and difference caption generation.
Anthology ID:
2025.acl-long.1428
Volume:
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
29503–29530
Language:
URL:
https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1428/
DOI:
Bibkey:
Cite (ACL):
Ron Yosef, Yonatan Bitton, Dani Lischinski, and Moran Yanuka. 2025. EditInspector: A Benchmark for Evaluation of Text-Guided Image Edits. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 29503–29530, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
EditInspector: A Benchmark for Evaluation of Text-Guided Image Edits (Yosef et al., ACL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.1428.pdf