Chi Kuan Lai
2025
NEKO at SemEval-2025 Task 4: A Gradient Ascent Based Machine Unlearning Strategy
Chi Kuan Lai
|
Yifei Chen
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
The power and wide application of large language models (LLMs) has brought the concerns on its risk of leaking private or sensitive information. However, retraining the modules is expensive and impractical, which introduces machine unlearning - removing specific information from language models while preserving general utility. Task 4 at SemEval 2025 consists of a shared task with this exact objective. We present an approach which combines gradient ascent-based forgetting with Kullback-Leibler (KL) divergence-based retention, applied to a 1-billion-parameter causal language model. Despite achieving effective forgetting, the system struggles with maintaining model utility. Our experiments reveal critical trade-off between unlearning effectiveness and performance preservation, highlighting challenges in practical machine unlearning implementations.