I2VWM: Robust Watermarking for Image to Video Generation
Guanjie Wang 1, Zehua Ma 1, Han Fang 2, Weiming Zhang 1
Published on arXiv
2509.17773
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
I2VWM significantly improves cross-modal watermark robustness across video frames while maintaining imperceptibility, outperforming single-modality watermarking baselines on I2V models.
I2VWM
Novel technique introduced
The rapid progress of image-guided video generation (I2V) has raised concerns about its potential misuse in misinformation and fraud, underscoring the urgent need for effective digital watermarking. While existing watermarking methods demonstrate robustness within a single modality, they fail to trace source images in I2V settings. To address this gap, we introduce the concept of Robust Diffusion Distance, which measures the temporal persistence of watermark signals in generated videos. Building on this, we propose I2VWM, a cross-modal watermarking framework designed to enhance watermark robustness across time. I2VWM leverages a video-simulation noise layer during training and employs an optical-flow-based alignment module during inference. Experiments on both open-source and commercial I2V models demonstrate that I2VWM significantly improves robustness while maintaining imperceptibility, establishing a new paradigm for cross-modal watermarking in the era of generative video. \href{https://github.com/MrCrims/I2VWM-Robust-Watermarking-for-Image-to-Video-Generation}{Code Released.}
Key Contributions
- Introduces Robust Diffusion Distance, a metric measuring temporal persistence of watermark signals across frames in I2V-generated videos
- Proposes I2VWM, a cross-modal watermarking framework using a video-simulation noise layer during training and optical-flow-based alignment at inference to maintain watermark robustness
- Demonstrates effectiveness on both open-source (Stable Video Diffusion, HunyuanVideo) and commercial I2V models
🛡️ Threat Analysis
Proposes content watermarking of AI-generated video outputs for provenance tracking — watermarks are embedded in source images to persist into generated videos, enabling traceability and authenticity verification of AI-generated content. This is output integrity and content provenance, not model IP protection.