benchmark 2025

AVFakeBench: A Comprehensive Audio-Video Forgery Detection Benchmark for AV-LMMs

Shuhan Xia 1, Peipei Li 1, Xuannan Liu 1, Dongsen Zhang 1, Xinyu Guo 2, Zekun Li 1

0 citations · 61 references · arXiv

α

Published on arXiv

2511.21251

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Evaluation of 11 AV-LMMs reveals promising detection potential but notable weaknesses in fine-grained forgery perception and reasoning tasks compared to specialized detection methods.

AVFakeBench

Novel technique introduced


The threat of Audio-Video (AV) forgery is rapidly evolving beyond human-centric deepfakes to include more diverse manipulations across complex natural scenes. However, existing benchmarks are still confined to DeepFake-based forgeries and single-granularity annotations, thus failing to capture the diversity and complexity of real-world forgery scenarios. To address this, we introduce AVFakeBench, the first comprehensive audio-video forgery detection benchmark that spans rich forgery semantics across both human subject and general subject. AVFakeBench comprises 12K carefully curated audio-video questions, covering seven forgery types and four levels of annotations. To ensure high-quality and diverse forgeries, we propose a multi-stage hybrid forgery framework that integrates proprietary models for task planning with expert generative models for precise manipulation. The benchmark establishes a multi-task evaluation framework covering binary judgment, forgery types classification, forgery detail selection, and explanatory reasoning. We evaluate 11 Audio-Video Large Language Models (AV-LMMs) and 2 prevalent detection methods on AVFakeBench, demonstrating the potential of AV-LMMs as emerging forgery detectors while revealing their notable weaknesses in fine-grained perception and reasoning.


Key Contributions

  • AVFakeBench: a 12K audio-video question benchmark spanning 7 forgery types (human and general subjects) with 4 annotation granularity levels
  • Multi-stage hybrid forgery framework integrating proprietary planning models with expert generative models for diverse, high-quality AV forgeries
  • Multi-task evaluation framework (binary judgment, forgery type classification, detail selection, explanatory reasoning) revealing fine-grained perception weaknesses in 11 AV-LMMs

🛡️ Threat Analysis

Output Integrity Attack

The benchmark directly targets AI-generated content detection — specifically audio-video forgeries and deepfakes. Creating evaluation frameworks, datasets, and metrics for detecting forged/synthetic multimodal content is a core ML09 contribution (output integrity and AI-generated content detection).


Details

Domains
visionaudiomultimodalnlp
Model Types
llmvlmmultimodal
Threat Tags
inference_time
Datasets
AVFakeBench
Applications
audio-video deepfake detectionmultimodal forgery detection