Deepfake Synthesis vs. Detection: An Uneven Contest
Md. Tarek Hasan 1, Sanjay Saha 2, Shaojing Fan 2, Swakkhar Shatabda 3, Terence Sim 2
Published on arXiv
2602.07986
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
State-of-the-art deepfake detectors exhibit markedly poor performance against deepfakes produced by modern synthesis techniques, with human participants also performing poorly against best-quality deepfakes
The rapid advancement of deepfake technology has significantly elevated the realism and accessibility of synthetic media. Emerging techniques, such as diffusion-based models and Neural Radiance Fields (NeRF), alongside enhancements in traditional Generative Adversarial Networks (GANs), have contributed to the sophisticated generation of deepfake videos. Concurrently, deepfake detection methods have seen notable progress, driven by innovations in Transformer architectures, contrastive learning, and other machine learning approaches. In this study, we conduct a comprehensive empirical analysis of state-of-the-art deepfake detection techniques, including human evaluation experiments against cutting-edge synthesis methods. Our findings highlight a concerning trend: many state-of-the-art detection models exhibit markedly poor performance when challenged with deepfakes produced by modern synthesis techniques, including poor performance by human participants against the best quality deepfakes. Through extensive experimentation, we provide evidence that underscores the urgent need for continued refinement of detection models to keep pace with the evolving capabilities of deepfake generation technologies. This research emphasizes the critical gap between current detection methodologies and the sophistication of new generation techniques, calling for intensified efforts in this crucial area of study.
Key Contributions
- Comprehensive empirical evaluation of state-of-the-art deepfake detection methods against modern synthesis techniques including diffusion models and NeRF
- Human evaluation experiments measuring human ability to detect high-quality deepfakes from cutting-edge generators
- Quantitative evidence of a critical performance gap between current detectors and modern deepfake generation methods
🛡️ Threat Analysis
Empirically evaluates AI-generated content detection (deepfake video detection) against modern synthesis techniques including diffusion models, NeRF, and GANs — directly addresses output integrity and authenticity of synthetic media.