Rahee Walambe
2025
NeuroReset : LLM Unlearning via Dual Phase Mixed Methodology
Dhwani Bhavankar
|
Het Sevalia
|
Shubh Agarwal
|
Yogesh Kulkarni
|
Rahee Walambe
|
Ketan Kotecha
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
This paper presents the method for the unlearning of sensitive information from large language models as applied in the SemEval 2025 Task 4 challenge. The unlearning pipeline consists of two phases. In phase I, the model is instructed to forget specific datasets, and in phase II, the model is stabilized using a retention dataset. Unlearning with these methods secured a final score of 0.420 with the 2nd honorary mention in the 7B parameter challenge and a score of 0.36 in the 13th position for the 1B parameter challenge. The paper presents a background study, a brief literature review, and a gap analysis, as well as the methodology employed in our work titled NeuroReset. The training methodology and evaluation metrics are also presented, and the trade-offs between unlearning efficiency and model performance are discussed. The contributions of the paper are systematic unlearning, a comparative analysis of unlearning methods, and an empirical analysis of model performance post-unlearning.