defense 2026

Multi-Feature Fusion Approach for Generative AI Images Detection

Abderrezzaq Sendjasni , Mohamed-Chaker Larabi

0 citations

α

Published on arXiv

2603.29788

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Fusion of statistical, semantic, and texture features yields superior and more consistent performance compared to state-of-the-art single-feature methods across all evaluated datasets, particularly in challenging mixed-model scenarios

Multi-Feature Fusion Framework

Novel technique introduced


The rapid evolution of Generative AI (GenAI) models has led to synthetic images of unprecedented realism, challenging traditional methods for distinguishing them from natural photographs. While existing detectors often rely on single-feature spaces, such as statistical regularities, semantic embeddings, or texture patterns, these approaches tend to lack robustness when confronted with diverse and evolving generative models. In this work, we investigate and systematically evaluate a multi-feature fusion framework that combines complementary cues from three distinct spaces: (1) Mean Subtracted Contrast Normalized (MSCN) features capturing low-level statistical deviations; (2) CLIP embeddings encoding high-level semantic coherence; and (3) Multi-scale Local Binary Patterns (MLBP) characterizing mid-level texture anomalies. Through extensive experiments on four benchmark datasets covering a wide range of generative models, we show that individual feature spaces exhibit significant performance variability across different generators. Crucially, the fusion of all three representations yields superior and more consistent performance, particularly in a challenging mixed-model scenario. Compared to state-of-the-art methods, the proposed framework yields consistently improved performance across all evaluated datasets. Overall, this work highlights the importance of hybrid representations for robust GenAI image detection and provides a principled framework for integrating complementary visual cues.


Key Contributions

  • Multi-feature fusion framework combining MSCN (statistical), CLIP (semantic), and MLBP (texture) features for GenAI image detection
  • Comprehensive benchmarking across four datasets showing fusion consistently outperforms single-feature approaches
  • Empirical demonstration of feature complementarity through dimensionality reduction analysis and mixed-model evaluation

🛡️ Threat Analysis

Output Integrity Attack

Detects AI-generated images (synthetic content from StyleGAN, DALL-E, Stable Diffusion, Midjourney) to verify output authenticity and distinguish natural from synthetic content — core output integrity and AI-generated content detection task.


Details

Domains
visiongenerative
Model Types
gandiffusion
Threat Tags
inference_time
Applications
ai-generated image detectiondeepfake detectionmultimedia forensics