defense 2025

CLIP-Flow: A Universal Discriminator for AI-Generated Images Inspired by Anomaly Detection

Zhipeng Yuan 1, Kai Wang 2, Weize Quan 3, Dong-Ming Yan 3, Tieru Wu 1

0 citations

α

Published on arXiv

2508.09477

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

CLIP-Flow generalizes effectively to AIIs from diverse unseen generative models (GANs, diffusion models) without requiring any AI-generated images at training time

CLIP-Flow

Novel technique introduced


With the rapid advancement of AI generative models, the visual quality of AI-generated images (AIIs) has become increasingly close to natural images, which inevitably raises security concerns. Most AII detectors often employ the conventional image classification pipeline with natural images and AIIs (generated by a generative model), which can result in limited detection performance for AIIs from unseen generative models. To solve this, we proposed a universal AI-generated image detector from the perspective of anomaly detection. Our discriminator does not need to access any AIIs and learn a generalizable representation with unsupervised learning. Specifically, we use the pre-trained CLIP encoder as the feature extractor and design a normalizing flow-like unsupervised model. Instead of AIIs, proxy images, e.g., obtained by applying a spectral modification operation on natural images, are used for training. Our models are trained by minimizing the likelihood of proxy images, optionally combined with maximizing the likelihood of natural images. Extensive experiments demonstrate the effectiveness of our method on AIIs produced by various image generators.


Key Contributions

  • Frames AI-generated image detection as an anomaly detection problem, eliminating the need for AI-generated images during training and improving generalization to unseen generators
  • Combines a pre-trained CLIP encoder with a normalizing flow model trained to maximize likelihood of natural images and minimize likelihood of proxy images
  • Introduces spectral modification-based proxy image generation as a substitute for AI-generated images in unsupervised training

🛡️ Threat Analysis

Output Integrity Attack

Proposes a novel AI-generated image detector — a direct output integrity / content authenticity contribution. The detector generalizes to unseen generative models by framing AII detection as anomaly detection over natural image distributions, using a CLIP encoder and normalizing flow trained only on natural and proxy images.


Details

Domains
vision
Model Types
transformergandiffusion
Threat Tags
inference_timeblack_box
Datasets
ForenSynthsGenImage
Applications
ai-generated image detectionimage forensics