defense 2026

Anti-I2V: Safeguarding your photos from malicious image-to-video generation

Duc Vu , Anh Nguyen , Chi Tran , Anh Tran

0 citations

α

Published on arXiv

2603.24570

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Achieves state-of-the-art defense performance against diverse video diffusion models by disrupting temporal coherence and generation fidelity

Anti-I2V

Novel technique introduced


Advances in diffusion-based video generation models, while significantly improving human animation, poses threats of misuse through the creation of fake videos from a specific person's photo and text prompts. Recent efforts have focused on adversarial attacks that introduce crafted perturbations to protect images from diffusion-based models. However, most existing approaches target image generation, while relatively few explicitly address image-to-video diffusion models (VDMs), and most primarily focus on UNet-based architectures. Hence, their effectiveness against Diffusion Transformer (DiT) models remains largely under-explored, as these models demonstrate improved feature retention, and stronger temporal consistency due to larger capacity and advanced attention mechanisms. In this work, we introduce Anti-I2V, a novel defense against malicious human image-to-video generation, applicable across diverse diffusion backbones. Instead of restricting noise updates to the RGB space, Anti-I2V operates in both the $L$*$a$*$b$* and frequency domains, improving robustness and concentrating on salient pixels. We then identify the network layers that capture the most distinct semantic features during the denoising process to design appropriate training objectives that maximize degradation of temporal coherence and generation fidelity. Through extensive validation, Anti-I2V demonstrates state-of-the-art defense performance against diverse video diffusion models, offering an effective solution to the problem.


Key Contributions

  • Defense operating in L*a*b* and frequency domains to improve adversarial robustness against video generation
  • Layer-specific attack objectives targeting temporal coherence degradation in DiT-based video diffusion models
  • State-of-the-art protection against diverse image-to-video generation models including Diffusion Transformer architectures

🛡️ Threat Analysis

Input Manipulation Attack

Applies adversarial perturbations to images at inference time to cause image-to-video models to generate degraded outputs — this is an adversarial input manipulation defense against malicious use of generative models.


Details

Domains
visionmultimodalgenerative
Model Types
diffusiontransformer
Threat Tags
inference_timedigital
Applications
image-to-video generationhuman animationdeepfake prevention