NEKO at SemEval-2025 Task 4: A Gradient Ascent Based Machine Unlearning Strategy

Chi Kuan Lai, Yifei Chen


Abstract
The power and wide application of large language models (LLMs) has brought the concerns on its risk of leaking private or sensitive information. However, retraining the modules is expensive and impractical, which introduces machine unlearning - removing specific information from language models while preserving general utility. Task 4 at SemEval 2025 consists of a shared task with this exact objective. We present an approach which combines gradient ascent-based forgetting with Kullback-Leibler (KL) divergence-based retention, applied to a 1-billion-parameter causal language model. Despite achieving effective forgetting, the system struggles with maintaining model utility. Our experiments reveal critical trade-off between unlearning effectiveness and performance preservation, highlighting challenges in practical machine unlearning implementations.
Anthology ID:
2025.semeval-1.64
Volume:
Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Sara Rosenthal, Aiala Rosá, Debanjan Ghosh, Marcos Zampieri
Venues:
SemEval | WS
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
463–467
Language:
URL:
https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.64/
DOI:
Bibkey:
Cite (ACL):
Chi Kuan Lai and Yifei Chen. 2025. NEKO at SemEval-2025 Task 4: A Gradient Ascent Based Machine Unlearning Strategy. In Proceedings of the 19th International Workshop on Semantic Evaluation (SemEval-2025), pages 463–467, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
NEKO at SemEval-2025 Task 4: A Gradient Ascent Based Machine Unlearning Strategy (Lai & Chen, SemEval 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/corrections-2025-08/2025.semeval-1.64.pdf