Semantic-Aware Reconstruction Error for Detecting AI-Generated Images
Ju Yeon Kang 1, Jaehong Park 1, Semin Kim 1, Ji Won Yoon 2, Nam Soo Kim 1
Published on arXiv
2508.09487
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
SARE outperforms existing baselines on GenImage and ForenSynths by exploiting the semantic gap between real images and their caption-guided reconstructions, achieving strong generalization to unseen generative models.
SARE (Semantic-Aware Reconstruction Error)
Novel technique introduced
Recently, AI-generated image detection has gained increasing attention, as the rapid advancement of image generation technologies has raised serious concerns about their potential misuse. While existing detection methods have achieved promising results, their performance often degrades significantly when facing fake images from unseen, out-of-distribution (OOD) generative models, since they primarily rely on model-specific artifacts and thus overfit to the models used for training. To address this limitation, we propose a novel representation, namely Semantic-Aware Reconstruction Error (SARE), that measures the semantic difference between an image and its caption-guided reconstruction. The key hypothesis behind SARE is that real images, whose captions often fail to fully capture their complex visual content, may undergo noticeable semantic shifts during the caption-guided reconstruction process. In contrast, fake images, which closely align with their captions, show minimal semantic changes. By quantifying these semantic shifts, SARE provides a robust and discriminative feature for detecting fake images across diverse generative models. Additionally, we introduce a fusion module that integrates SARE into the backbone detector via a cross-attention mechanism. Image features attend to semantic representations extracted from SARE, enabling the model to adaptively leverage semantic information. Experimental results demonstrate that the proposed method achieves strong generalization, outperforming existing baselines on benchmarks including GenImage and ForenSynths. We further validate the effectiveness of caption guidance through a detailed analysis of semantic shifts, confirming its ability to enhance detection robustness.
Key Contributions
- Semantic-Aware Reconstruction Error (SARE): a novel feature that quantifies semantic shift between an image and its caption-guided reconstruction to distinguish real from AI-generated images.
- Cross-attention fusion module that integrates SARE representations into a backbone detector, enabling adaptive use of semantic information.
- Demonstrated strong out-of-distribution generalization over unseen generative models on GenImage and ForenSynths benchmarks, outperforming existing baselines.
🛡️ Threat Analysis
Proposes a novel forensic detection technique for AI-generated images by measuring semantic shifts during caption-guided reconstruction — directly addresses output integrity and content authenticity for images from diverse generative models.