defense 2025

Semantic-Aware Reconstruction Error for Detecting AI-Generated Images

Ju Yeon Kang 1, Jaehong Park 1, Semin Kim 1, Ji Won Yoon 2, Nam Soo Kim 1

0 citations

α

Published on arXiv

2508.09487

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

SARE outperforms existing baselines on GenImage and ForenSynths by exploiting the semantic gap between real images and their caption-guided reconstructions, achieving strong generalization to unseen generative models.

SARE (Semantic-Aware Reconstruction Error)

Novel technique introduced


Recently, AI-generated image detection has gained increasing attention, as the rapid advancement of image generation technologies has raised serious concerns about their potential misuse. While existing detection methods have achieved promising results, their performance often degrades significantly when facing fake images from unseen, out-of-distribution (OOD) generative models, since they primarily rely on model-specific artifacts and thus overfit to the models used for training. To address this limitation, we propose a novel representation, namely Semantic-Aware Reconstruction Error (SARE), that measures the semantic difference between an image and its caption-guided reconstruction. The key hypothesis behind SARE is that real images, whose captions often fail to fully capture their complex visual content, may undergo noticeable semantic shifts during the caption-guided reconstruction process. In contrast, fake images, which closely align with their captions, show minimal semantic changes. By quantifying these semantic shifts, SARE provides a robust and discriminative feature for detecting fake images across diverse generative models. Additionally, we introduce a fusion module that integrates SARE into the backbone detector via a cross-attention mechanism. Image features attend to semantic representations extracted from SARE, enabling the model to adaptively leverage semantic information. Experimental results demonstrate that the proposed method achieves strong generalization, outperforming existing baselines on benchmarks including GenImage and ForenSynths. We further validate the effectiveness of caption guidance through a detailed analysis of semantic shifts, confirming its ability to enhance detection robustness.


Key Contributions

  • Semantic-Aware Reconstruction Error (SARE): a novel feature that quantifies semantic shift between an image and its caption-guided reconstruction to distinguish real from AI-generated images.
  • Cross-attention fusion module that integrates SARE representations into a backbone detector, enabling adaptive use of semantic information.
  • Demonstrated strong out-of-distribution generalization over unseen generative models on GenImage and ForenSynths benchmarks, outperforming existing baselines.

🛡️ Threat Analysis

Output Integrity Attack

Proposes a novel forensic detection technique for AI-generated images by measuring semantic shifts during caption-guided reconstruction — directly addresses output integrity and content authenticity for images from diverse generative models.


Details

Domains
visiongenerative
Model Types
diffusiongantransformer
Threat Tags
inference_time
Datasets
GenImageForenSynths
Applications
ai-generated image detectiondeepfake detection