Towards a Principled Evaluation of Knowledge Editors

Sebastian Pohl, Max Ploner, Alan Akbik


Abstract
Model editing has been gaining increasing attention over the past few years. For Knowledge Editing in particular, more challenging evaluation datasets have recently been released. These datasets use different methodologies to score the success of editors. Yet, it remains under-explored how robust these methodologies are and whether they unfairly favor some editors. Moreover, the disruptive impact of these editors on overall model capabilities remains a constant blind spot.We address both of these problems and show that choosing different metrics and evaluation methodologies as well as different edit batch sizes can lead to a different ranking of knowledge editors. Crucially we demonstrate this effect also on general language understanding tasks evaluated alongside the knowledge editing tasks. Further we include a manual assessment of the string matching based evaluation method for knowledge editing that is favored by recently released datasets, revealing a tendency to produce false positive matches.
Anthology ID:
2025.l2m2-1.4
Volume:
Proceedings of the First Workshop on Large Language Model Memorization (L2M2)
Month:
August
Year:
2025
Address:
Vienna, Austria
Editors:
Robin Jia, Eric Wallace, Yangsibo Huang, Tiago Pimentel, Pratyush Maini, Verna Dankers, Johnny Wei, Pietro Lesci
Venues:
L2M2 | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
47–60
Language:
URL:
https://preview.aclanthology.org/landing_page/2025.l2m2-1.4/
DOI:
Bibkey:
Cite (ACL):
Sebastian Pohl, Max Ploner, and Alan Akbik. 2025. Towards a Principled Evaluation of Knowledge Editors. In Proceedings of the First Workshop on Large Language Model Memorization (L2M2), pages 47–60, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Towards a Principled Evaluation of Knowledge Editors (Pohl et al., L2M2 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/landing_page/2025.l2m2-1.4.pdf