NS-Net: Decoupling CLIP Semantic Information through NULL-Space for Generalizable AI-Generated Image Detection
Jiazhen Yan 1, Fan Wang 2, Weiwei Jiang 1, Ziqiang Li 1, Zhangjie Fu 1
Published on arXiv
2508.01248
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
NS-Net achieves a 7.4% improvement in detection accuracy over state-of-the-art methods on an open-world benchmark spanning 40 diverse GAN- and diffusion-based generative models.
NS-Net
Novel technique introduced
The rapid progress of generative models, such as GANs and diffusion models, has facilitated the creation of highly realistic images, raising growing concerns over their misuse in security-sensitive domains. While existing detectors perform well under known generative settings, they often fail to generalize to unknown generative models, especially when semantic content between real and fake images is closely aligned. In this paper, we revisit the use of CLIP features for AI-generated image detection and uncover a critical limitation: the high-level semantic information embedded in CLIP's visual features hinders effective discrimination. To address this, we propose NS-Net, a novel detection framework that leverages NULL-Space projection to decouple semantic information from CLIP's visual features, followed by contrastive learning to capture intrinsic distributional differences between real and generated images. Furthermore, we design a Patch Selection strategy to preserve fine-grained artifacts by mitigating semantic bias caused by global image structures. Extensive experiments on an open-world benchmark comprising images generated by 40 diverse generative models show that NS-Net outperforms existing state-of-the-art methods, achieving a 7.4\% improvement in detection accuracy, thereby demonstrating strong generalization across both GAN- and diffusion-based image generation techniques.
Key Contributions
- NULL-Space projection to decouple high-level semantic information from CLIP visual features, reducing semantic bias in AI-generated image detection
- Contrastive learning module that captures intrinsic distributional differences between real and AI-generated images after semantic decoupling
- Patch Selection strategy based on spectral entropy to preserve fine-grained generation artifacts and mitigate global structure bias
🛡️ Threat Analysis
The paper's primary contribution is a novel detector for AI-generated images (from GANs and diffusion models), directly addressing output integrity and content authenticity — a canonical ML09 concern of AI-generated content detection.