benchmark 2026

How Effective Are Publicly Accessible Deepfake Detection Tools? A Comparative Evaluation of Open-Source and Free-to-Use Platforms

Michael Rettinger , Ben Beaumont , Nhien-An Le-Khac , Hong-Hanh Nguyen-Le

0 citations

α

Published on arXiv

2603.04456

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Human evaluators substantially outperform all six automated tools, and in cases of human-AI disagreement, human judgment prevails in the vast majority of discordant cases


The proliferation of deepfake imagery poses escalating challenges for practitioners tasked with verifying digital media authenticity. While detection algorithm research is abundant, empirical evaluations of publicly accessible tools that practitioners actually use remain scarce. This paper presents the first cross-paradigm evaluation of six tools, spanning two complementary detection approaches: forensic analysis tools (InVID \& WeVerify, FotoForensics, Forensically) and AI-based classifiers (DecopyAI, FaceOnLive, Bitmind). Both tool categories were evaluated by professional investigators with law enforcement experience using blinded protocols across datasets comprising authentic, tampered, and AI-generated images sourced from DF40, CelebDF, and CASIA-v2. We report three principal findings: forensic tools exhibit high recall but poor specificity, while AI classifiers demonstrate the inverse pattern; human evaluators substantially outperform all automated tools; and human-AI disagreement is asymmetric, with human judgment prevailing in the vast majority of discordant cases. We discuss implications for practitioner workflows and identify critical gaps in current detection capabilities.


Key Contributions

  • First cross-paradigm empirical evaluation of six publicly accessible deepfake detection tools (forensic platforms vs. AI classifiers) using professional law enforcement investigators
  • Identifies asymmetric performance trade-off: forensic tools show high recall/low specificity while AI classifiers show the inverse pattern
  • Demonstrates that human evaluators substantially outperform all automated tools and that human judgment prevails in the majority of human-AI disagreement cases

🛡️ Threat Analysis

Output Integrity Attack

Paper evaluates AI-generated image detection tools (deepfake classifiers and forensic platforms) — directly addresses output integrity and AI-generated content detection, which is a core ML09 concern.


Details

Domains
visiongenerative
Model Types
cnntransformer
Threat Tags
inference_time
Datasets
DF40CelebDFCASIA-v2
Applications
deepfake detectiondigital media forensicslaw enforcement image verification