@inproceedings{zhu-etal-2025-initializing,
title = "Initializing and Retrofitting Key-Value Adaptors for Traceable Model Editing",
author = "Zhu, Hanlun and
Lan, Yunshi and
Li, Xiang and
Qian, Weining",
editor = "Che, Wanxiang and
Nabende, Joyce and
Shutova, Ekaterina and
Pilehvar, Mohammad Taher",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2025",
month = jul,
year = "2025",
address = "Vienna, Austria",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.152/",
pages = "2958--2971",
ISBN = "979-8-89176-256-5",
abstract = "As the insight of knowledge storage in language models deepens, the ability to perform CRUD (Create, Read, Update, Delete) operations on language models becomes increasingly indispensable for satisfying the demands of managing rapidly updating knowledge. Considering the high cost of fine-tuning language models, model editing methods with low cost are usually required to manipulate models' knowledge. The evidence suggests that modules carrying knowledge in a Transformer module are primarily the MLP blocks, thus we propose iReVa, a method that explicitly initializes and retrofits key-value pairs into MLP blocks to construct a new mapping of a piece of knowledge without damaging the irrelevant knowledge. In comparison to existing methods, iReVa reveals better interpretability and a stronger capacity for carrying traceable edits. Experiment results on a series of GPT series models show our prominent performance on edit success and generalization without influencing specificity. We also made the first attempt to conduct a knowledge withdrawal test of iReVa. Our codes are available at https://github.com/timberflow/iReVa."
}
Markdown (Informal)
[Initializing and Retrofitting Key-Value Adaptors for Traceable Model Editing](https://preview.aclanthology.org/ingestion-acl-25/2025.findings-acl.152/) (Zhu et al., Findings 2025)
ACL