defense 2025

I2VWM: Robust Watermarking for Image to Video Generation

Guanjie Wang 1, Zehua Ma 1, Han Fang 2, Weiming Zhang 1

0 citations · 31 references · arXiv

α

Published on arXiv

2509.17773

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

I2VWM significantly improves cross-modal watermark robustness across video frames while maintaining imperceptibility, outperforming single-modality watermarking baselines on I2V models.

I2VWM

Novel technique introduced


The rapid progress of image-guided video generation (I2V) has raised concerns about its potential misuse in misinformation and fraud, underscoring the urgent need for effective digital watermarking. While existing watermarking methods demonstrate robustness within a single modality, they fail to trace source images in I2V settings. To address this gap, we introduce the concept of Robust Diffusion Distance, which measures the temporal persistence of watermark signals in generated videos. Building on this, we propose I2VWM, a cross-modal watermarking framework designed to enhance watermark robustness across time. I2VWM leverages a video-simulation noise layer during training and employs an optical-flow-based alignment module during inference. Experiments on both open-source and commercial I2V models demonstrate that I2VWM significantly improves robustness while maintaining imperceptibility, establishing a new paradigm for cross-modal watermarking in the era of generative video. \href{https://github.com/MrCrims/I2VWM-Robust-Watermarking-for-Image-to-Video-Generation}{Code Released.}


Key Contributions

  • Introduces Robust Diffusion Distance, a metric measuring temporal persistence of watermark signals across frames in I2V-generated videos
  • Proposes I2VWM, a cross-modal watermarking framework using a video-simulation noise layer during training and optical-flow-based alignment at inference to maintain watermark robustness
  • Demonstrates effectiveness on both open-source (Stable Video Diffusion, HunyuanVideo) and commercial I2V models

🛡️ Threat Analysis

Output Integrity Attack

Proposes content watermarking of AI-generated video outputs for provenance tracking — watermarks are embedded in source images to persist into generated videos, enabling traceability and authenticity verification of AI-generated content. This is output integrity and content provenance, not model IP protection.


Details

Domains
visiongenerative
Model Types
diffusion
Threat Tags
inference_time
Datasets
MS-COCODIV2K (NTIRE 2017)
Applications
image-to-video generationvideo content provenancedeepfake/misuse tracing