defense 2025

NS-Net: Decoupling CLIP Semantic Information through NULL-Space for Generalizable AI-Generated Image Detection

Jiazhen Yan 1, Fan Wang 2, Weiwei Jiang 1, Ziqiang Li 1, Zhangjie Fu 1

0 citations

α

Published on arXiv

2508.01248

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

NS-Net achieves a 7.4% improvement in detection accuracy over state-of-the-art methods on an open-world benchmark spanning 40 diverse GAN- and diffusion-based generative models.

NS-Net

Novel technique introduced


The rapid progress of generative models, such as GANs and diffusion models, has facilitated the creation of highly realistic images, raising growing concerns over their misuse in security-sensitive domains. While existing detectors perform well under known generative settings, they often fail to generalize to unknown generative models, especially when semantic content between real and fake images is closely aligned. In this paper, we revisit the use of CLIP features for AI-generated image detection and uncover a critical limitation: the high-level semantic information embedded in CLIP's visual features hinders effective discrimination. To address this, we propose NS-Net, a novel detection framework that leverages NULL-Space projection to decouple semantic information from CLIP's visual features, followed by contrastive learning to capture intrinsic distributional differences between real and generated images. Furthermore, we design a Patch Selection strategy to preserve fine-grained artifacts by mitigating semantic bias caused by global image structures. Extensive experiments on an open-world benchmark comprising images generated by 40 diverse generative models show that NS-Net outperforms existing state-of-the-art methods, achieving a 7.4\% improvement in detection accuracy, thereby demonstrating strong generalization across both GAN- and diffusion-based image generation techniques.


Key Contributions

  • NULL-Space projection to decouple high-level semantic information from CLIP visual features, reducing semantic bias in AI-generated image detection
  • Contrastive learning module that captures intrinsic distributional differences between real and AI-generated images after semantic decoupling
  • Patch Selection strategy based on spectral entropy to preserve fine-grained generation artifacts and mitigate global structure bias

🛡️ Threat Analysis

Output Integrity Attack

The paper's primary contribution is a novel detector for AI-generated images (from GANs and diffusion models), directly addressing output integrity and content authenticity — a canonical ML09 concern of AI-generated content detection.


Details

Domains
visiongenerative
Model Types
transformergandiffusion
Threat Tags
inference_time
Datasets
LSUNopen-world benchmark (40 generative models)
Applications
ai-generated image detectiondeepfake detection