defense 2026

Layer Consistency Matters: Elegant Latent Transition Discrepancy for Generalizable Synthetic Image Detection

Yawen Yang , Feng Li , Shuqi Kong , Yunfeng Diao , Xinjian Gao , Zenglin Shi , Meng Wang

0 citations

α

Published on arXiv

2603.10598

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

LTD exceeds the base model by 14.35% in mean accuracy across three datasets covering diverse GAN and diffusion model generators, achieving superior generalizability and robustness over recent state-of-the-art methods

Latent Transition Discrepancy (LTD)

Novel technique introduced


Recent rapid advancement of generative models has significantly improved the fidelity and accessibility of AI-generated synthetic images. While enabling various innovative applications, the unprecedented realism of these synthetics makes them increasingly indistinguishable from authentic photographs, posing serious security risks, such as media credibility and content manipulation. Although extensive efforts have been dedicated to detecting synthetic images, most existing approaches suffer from poor generalization to unseen data due to their reliance on model-specific artifacts or low-level statistical cues. In this work, we identify a previously unexplored distinction that real images maintain consistent semantic attention and structural coherence in their latent representations, exhibiting more stable feature transitions across network layers, whereas synthetic ones present discernible distinct patterns. Therefore, we propose a novel approach termed latent transition discrepancy (LTD), which captures the inter-layer consistency differences of real and synthetic images. LTD adaptively identifies the most discriminative layers and assesses the transition discrepancies across layers. Benefiting from the proposed inter-layer discriminative modeling, our approach exceeds the base model by 14.35\% in mean Acc across three datasets containing diverse GANs and DMs. Extensive experiments demonstrate that LTD outperforms recent state-of-the-art methods, achieving superior detection accuracy, generalizability, and robustness. The code is available at https://github.com/yywencs/LTD


Key Contributions

  • Identifies that real images exhibit more stable feature transitions across network layers compared to synthetic images from GANs and diffusion models
  • Proposes Latent Transition Discrepancy (LTD), which adaptively selects the most discriminative layers and measures inter-layer consistency differences to distinguish real from synthetic images
  • Achieves 14.35% improvement in mean accuracy over the base model across three diverse benchmarks, outperforming state-of-the-art synthetic image detectors

🛡️ Threat Analysis

Output Integrity Attack

Proposes a novel synthetic image detector that distinguishes real from AI-generated images (GANs and diffusion models) — directly addresses output integrity and AI-generated content detection.


Details

Domains
visiongenerative
Model Types
gandiffusioncnntransformer
Threat Tags
inference_time
Datasets
ForenSynthsGenImageUniversalFakeDetect
Applications
synthetic image detectiondeepfake detectionmedia integrity verification