defense 2026

Generalizable Detection of AI Generated Images with Large Models and Fuzzy Decision Tree

Fei Wu , Guanghao Ding , Zijian Niu , Zhenrui Wang , Lei Yang , Zhuosheng Zhang , Shilin Wang

0 citations

α

Published on arXiv

2603.28508

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Achieves state-of-the-art detection accuracy with strong generalization across diverse generative models by fusing MLLM semantic reasoning with artifact-aware perceptual detectors

Fuzzy Decision Tree AIGC Detector

Novel technique introduced


The malicious use and widespread dissemination of AI-generated images pose a serious threat to the authenticity of digital content. Existing detection methods exploit low-level artifacts left by common manipulation steps within the generation pipeline, but they often lack generalization due to model-specific overfitting. Recently, researchers have resorted to Multimodal Large Language Models (MLLMs) for AIGC detection, leveraging their high-level semantic reasoning and broad generalization capabilities. While promising, MLLMs lack the fine-grained perceptual sensitivity to subtle generation artifacts, making them inadequate as standalone detectors. To address this issue, we propose a novel AI-generated image detection framework that synergistically integrates lightweight artifact-aware detectors with MLLMs via a fuzzy decision tree. The decision tree treats the outputs of basic detectors as fuzzy membership values, enabling adaptive fusion of complementary cues from semantic and perceptual perspectives. Extensive experiments demonstrate that the proposed method achieves state-of-the-art accuracy and strong generalization across diverse generative models.


Key Contributions

  • Novel framework integrating lightweight artifact-aware detectors with MLLMs via fuzzy decision tree
  • Adaptive fusion of complementary cues from semantic (MLLM) and perceptual (artifact) perspectives
  • State-of-the-art accuracy with strong generalization across diverse generative models

🛡️ Threat Analysis

Output Integrity Attack

Detects AI-generated images to verify content authenticity and provenance — this is output integrity verification, specifically detecting synthetic/fake content.


Details

Domains
visionmultimodal
Model Types
multimodalgandiffusion
Threat Tags
inference_time
Applications
ai-generated image detectioncontent authenticity verificationdeepfake detection