Naen Xu
2025
VideoEraser: Concept Erasure in Text-to-Video Diffusion Models
Naen Xu
|
Jinghuai Zhang
|
Changjiang Li
|
Zhi Chen
|
Chunyi Zhou
|
Qingming Li
|
Tianyu Du
|
Shouling Ji
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
The rapid growth of text-to-video (T2V) diffusion models has raised concerns about privacy, copyright, and safety due to their potential misuse in generating harmful or misleading content. These models are often trained on numerous datasets, including unauthorized personal identities, artistic creations, and harmful materials, which can lead to uncontrolled production and distribution of such content. To address this, we propose VideoEraser, a training-free framework that prevents T2V diffusion models from generating videos with undesirable concepts, even when explicitly prompted with those concepts. Designed as a plug-and-play module, VideoEraser can seamlessly integrate with representative T2V diffusion models via a two-stage process: Selective Prompt Embedding Adjustment (SPEA) and Adversarial-Resilient Noise Guidance (ARNG). We conduct extensive evaluations across four tasks, including object erasure, artistic style erasure, celebrity erasure, and explicit content erasure. Experimental results show that VideoEraser consistently outperforms prior methods regarding efficacy, integrity, fidelity, robustness, and generalizability. Notably, VideoEraser achieves state-of-the-art performance in suppressing undesirable content during T2V generation, reducing it by 46% on average across four tasks compared to baselines.
Search
Fix author
Co-authors
- Zhi Chen 1
- Tianyu Du 1
- Shouling Ji 1
- Changjiang Li 1
- Qingming Li 1
- show all...