TriDF: Evaluating Perception, Detection, and Hallucination for Interpretable DeepFake Detection
Jian-Yu Jiang-Lin 1, Kang-Yang Huang 1, Ling Zou 1, Ling Lo 2, Sheng-Ping Yang 1, Yu-Wen Tseng 1, Kun-Hsiang Lin 1, Chia-Ling Chen 1, Yu-Ting Ta 1, Yanting Wang 1, Po-Ching Chen 1, Hongxia Xie 3, Hong-Han Shuai 2, Wen-Huang Cheng 1
Published on arXiv
2512.10652
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
Experiments on state-of-the-art MLLMs reveal that accurate perception is essential for reliable detection, but hallucination in model explanations can severely disrupt decision-making, highlighting the interdependence of these three evaluation dimensions.
TriDF
Novel technique introduced
Advances in generative modeling have made it increasingly easy to fabricate realistic portrayals of individuals, creating serious risks for security, communication, and public trust. Detecting such person-driven manipulations requires systems that not only distinguish altered content from authentic media but also provide clear and reliable reasoning. In this paper, we introduce TriDF, a comprehensive benchmark for interpretable DeepFake detection. TriDF contains high-quality forgeries from advanced synthesis models, covering 16 DeepFake types across image, video, and audio modalities. The benchmark evaluates three key aspects: Perception, which measures the ability of a model to identify fine-grained manipulation artifacts using human-annotated evidence; Detection, which assesses classification performance across diverse forgery families and generators; and Hallucination, which quantifies the reliability of model-generated explanations. Experiments on state-of-the-art multimodal large language models show that accurate perception is essential for reliable detection, but hallucination can severely disrupt decision-making, revealing the interdependence of these three aspects. TriDF provides a unified framework for understanding the interaction between detection accuracy, evidence identification, and explanation reliability, offering a foundation for building trustworthy systems that address real-world synthetic media threats.
Key Contributions
- TriDF benchmark with 5K high-quality DeepFake samples spanning 16 manipulation types across image, video, and audio modalities
- Three-dimensional evaluation framework measuring Perception (artifact identification), Detection (classification), and Hallucination (explanation reliability) for interpretable deepfake detectors
- Empirical finding that hallucination in MLLM-generated explanations can severely disrupt detection decision-making even when perceptual ability is adequate
🛡️ Threat Analysis
TriDF is a benchmark specifically designed to evaluate AI-generated content detection (deepfakes across image, video, and audio), directly targeting output integrity and content authenticity — a canonical ML09 application.