Generalizable Detection of AI Generated Images with Large Models and Fuzzy Decision Tree
Fei Wu , Guanghao Ding , Zijian Niu , Zhenrui Wang , Lei Yang , Zhuosheng Zhang , Shilin Wang
Published on arXiv
2603.28508
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
Achieves state-of-the-art detection accuracy with strong generalization across diverse generative models by fusing MLLM semantic reasoning with artifact-aware perceptual detectors
Fuzzy Decision Tree AIGC Detector
Novel technique introduced
The malicious use and widespread dissemination of AI-generated images pose a serious threat to the authenticity of digital content. Existing detection methods exploit low-level artifacts left by common manipulation steps within the generation pipeline, but they often lack generalization due to model-specific overfitting. Recently, researchers have resorted to Multimodal Large Language Models (MLLMs) for AIGC detection, leveraging their high-level semantic reasoning and broad generalization capabilities. While promising, MLLMs lack the fine-grained perceptual sensitivity to subtle generation artifacts, making them inadequate as standalone detectors. To address this issue, we propose a novel AI-generated image detection framework that synergistically integrates lightweight artifact-aware detectors with MLLMs via a fuzzy decision tree. The decision tree treats the outputs of basic detectors as fuzzy membership values, enabling adaptive fusion of complementary cues from semantic and perceptual perspectives. Extensive experiments demonstrate that the proposed method achieves state-of-the-art accuracy and strong generalization across diverse generative models.
Key Contributions
- Novel framework integrating lightweight artifact-aware detectors with MLLMs via fuzzy decision tree
- Adaptive fusion of complementary cues from semantic (MLLM) and perceptual (artifact) perspectives
- State-of-the-art accuracy with strong generalization across diverse generative models
🛡️ Threat Analysis
Detects AI-generated images to verify content authenticity and provenance — this is output integrity verification, specifically detecting synthetic/fake content.