Minxin Du
2025
OBLIVIATE: Robust and Practical Machine Unlearning for Large Language Models
Xiaoyu Xu
|
Minxin Du
|
Qingqing Ye
|
Haibo Hu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Large language models (LLMs) trained over extensive corpora risk memorizing sensitive, copyrighted, or toxic content. To address this, we propose OBLIVIATE, a robust unlearning framework that removes targeted data while preserving model utility. The framework follows a structured process: extracting target tokens, building retain sets, and fine-tuning with a tailored loss function comprising three components—masking, distillation, and world fact. Using low-rank adapters (LoRA) ensures efficiency without compromising unlearning quality. We conduct experiments on multiple datasets, including Harry Potter series, WMDP, and TOFU, using a comprehensive suite of metrics: forget quality (via a new document-level memorization score), model utility, and fluency. Results demonstrate its effectiveness in resisting membership inference attacks, minimizing the impact on retained data, and maintaining robustness across diverse scenarios.
2024
Machine Unlearning of Pre-trained Large Language Models
Jin Yao
|
Eli Chien
|
Minxin Du
|
Xinyao Niu
|
Tianhao Wang
|
Zezhou Cheng
|
Xiang Yue
Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
This study investigates the concept of the ‘right to be forgotten’ within the context of large language models (LLMs). We explore machine unlearning as a pivotal solution, with a focus on pre-trained models–a notably under-researched area. Our research delineates a comprehensive framework for machine unlearning in pre-trained LLMs, encompassing a critical analysis of seven diverse unlearning methods. Through rigorous evaluation using curated datasets from arXiv, books, and GitHub, we establish a robust benchmark for unlearning performance, demonstrating that these methods are over 105 times more computationally efficient than retraining. Our results show that integrating gradient ascent with gradient descent on in-distribution data improves hyperparameter robustness. We also provide detailed guidelines for efficient hyperparameter tuning in the unlearning process. Our findings advance the discourse on ethical AI practices, offering substantive insights into the mechanics of machine unlearning for pre-trained LLMs and underscoring the potential for responsible AI development.
2021
Differential Privacy for Text Analytics via Natural Text Sanitization
Xiang Yue
|
Minxin Du
|
Tianhao Wang
|
Yaliang Li
|
Huan Sun
|
Sherman S. M. Chow
Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021
Search
Fix author
Co-authors
- Tianhao Wang 2
- Xiang Yue 2
- Zezhou Cheng 1
- Eli Chien 1
- Sherman S. M. Chow 1
- show all...