benchmark 2026

Deepfake Synthesis vs. Detection: An Uneven Contest

Md. Tarek Hasan 1, Sanjay Saha 2, Shaojing Fan 2, Swakkhar Shatabda 3, Terence Sim 2

0 citations · 33 references · arXiv (Cornell University)

α

Published on arXiv

2602.07986

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

State-of-the-art deepfake detectors exhibit markedly poor performance against deepfakes produced by modern synthesis techniques, with human participants also performing poorly against best-quality deepfakes


The rapid advancement of deepfake technology has significantly elevated the realism and accessibility of synthetic media. Emerging techniques, such as diffusion-based models and Neural Radiance Fields (NeRF), alongside enhancements in traditional Generative Adversarial Networks (GANs), have contributed to the sophisticated generation of deepfake videos. Concurrently, deepfake detection methods have seen notable progress, driven by innovations in Transformer architectures, contrastive learning, and other machine learning approaches. In this study, we conduct a comprehensive empirical analysis of state-of-the-art deepfake detection techniques, including human evaluation experiments against cutting-edge synthesis methods. Our findings highlight a concerning trend: many state-of-the-art detection models exhibit markedly poor performance when challenged with deepfakes produced by modern synthesis techniques, including poor performance by human participants against the best quality deepfakes. Through extensive experimentation, we provide evidence that underscores the urgent need for continued refinement of detection models to keep pace with the evolving capabilities of deepfake generation technologies. This research emphasizes the critical gap between current detection methodologies and the sophistication of new generation techniques, calling for intensified efforts in this crucial area of study.


Key Contributions

  • Comprehensive empirical evaluation of state-of-the-art deepfake detection methods against modern synthesis techniques including diffusion models and NeRF
  • Human evaluation experiments measuring human ability to detect high-quality deepfakes from cutting-edge generators
  • Quantitative evidence of a critical performance gap between current detectors and modern deepfake generation methods

🛡️ Threat Analysis

Output Integrity Attack

Empirically evaluates AI-generated content detection (deepfake video detection) against modern synthesis techniques including diffusion models, NeRF, and GANs — directly addresses output integrity and authenticity of synthetic media.


Details

Domains
visiongenerative
Model Types
gandiffusiontransformercnn
Threat Tags
inference_timedigital
Applications
deepfake detectionsynthetic media authenticationvideo forensics