defense 2026

Gaussian Shannon: High-Precision Diffusion Model Watermarking Based on Communication

Yi Zhang , Hongbo Huang , Liang-Jie Zhang

0 citations

α

Published on arXiv

2603.26167

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Achieves state-of-the-art bit-level accuracy for watermark recovery across seven perturbation types while maintaining high true positive rate

Gaussian Shannon

Novel technique introduced


Diffusion models generate high-quality images but pose serious risks like copyright violation and disinformation. Watermarking is a key defense for tracing and authenticating AI-generated content. However, existing methods rely on threshold-based detection, which only supports fuzzy matching and cannot recover structured watermark data bit-exactly, making them unsuitable for offline verification or applications requiring lossless metadata (e.g., licensing instructions). To address this problem, in this paper, we propose Gaussian Shannon, a watermarking framework that treats the diffusion process as a noisy communication channel and enables both robust tracing and exact bit recovery. Our method embeds watermarks in the initial Gaussian noise without fine-tuning or quality loss. We identify two types of channel interference, namely local bit flips and global stochastic distortions, and design a cascaded defense combining error-correcting codes and majority voting. This ensures reliable end-to-end transmission of semantic payloads. Experiments across three Stable Diffusion variants and seven perturbation types show that Gaussian Shannon achieves state-of-the-art bit-level accuracy while maintaining a high true positive rate, enabling trustworthy rights attribution in real-world deployment. The source code have been made available at: https://github.com/Rambo-Yi/Gaussian-Shannon


Key Contributions

  • Communication-theory-based watermarking framework treating diffusion as noisy channel
  • Cascaded defense combining error-correcting codes and majority voting for bit-exact recovery
  • No-fine-tuning embedding in initial Gaussian noise with zero quality loss

🛡️ Threat Analysis

Output Integrity Attack

Watermarks AI-generated images (diffusion model outputs) for content provenance, authentication, and copyright tracing — this is output integrity, not model theft protection.


Details

Domains
visiongenerative
Model Types
diffusion
Threat Tags
inference_time
Datasets
Stable Diffusion variants
Applications
image generationcopyright protectioncontent authentication