The Illusion of Forgetting: Attack Unlearned Diffusion via Initial Latent Variable Optimization
Manyi Li 1, Yufan Liu 2, Lai Jiang 3, Bing Li 2, Yuming Li 4, Weiming Hu 2
Published on arXiv
2602.00175
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
IVO successfully reactivates NSFW 'dormant memories' in models treated with 8 different unlearning techniques, achieving superior attack success rates and strong semantic consistency, demonstrating that current unlearning-based defenses are fundamentally ineffective.
IVO (Initial Latent Variable Optimization)
Novel technique introduced
Although unlearning-based defenses claim to purge Not-Safe-For-Work (NSFW) concepts from diffusion models (DMs), we reveals that this "forgetting" is largely an illusion. Unlearning partially disrupts the mapping between linguistic symbols and the underlying knowledge, which remains intact as dormant memories. We find that the distributional discrepancy in the denoising process serves as a measurable indicator of how much of the mapping is retained, also reflecting the strength of unlearning. Inspired by this, we propose IVO (Initial Latent Variable Optimization), a concise and powerful attack framework that reactivates these dormant memories by reconstructing the broken mappings. Through Image Inversion}, Adversarial Optimization and Reused Attack, IVO optimizes initial latent variables to realign the noise distribution of unlearned models with their original unsafe states. Extensive experiments across 8 widely used unlearning techniques demonstrate that IVO achieves superior attack success rates and strong semantic consistency, exposing fundamental flaws in current defenses. The code is available at anonymous.4open.science/r/IVO/. Warning: This paper has unsafe images that may offend some readers.
Key Contributions
- Theoretical insight that unlearning in diffusion models does not erase unsafe knowledge but only disrupts linguistic-to-knowledge mappings, leaving 'dormant memories' intact
- IVO framework using Image Inversion, Adversarial Optimization, and Reused Attack to optimize initial latent variables and realign unlearned model noise distributions with unsafe states
- Empirical demonstration of superior attack success rates and semantic consistency across 8 widely used unlearning techniques, exposing fundamental flaws in current defenses
🛡️ Threat Analysis
IVO is a white-box adversarial attack that uses gradient-based optimization of initial latent variables (crafted inputs) at inference time to manipulate diffusion model outputs and bypass safety unlearning defenses — directly fitting the core definition of input manipulation via adversarial optimization.