defense 2026

Direct Discrepancy Replay: Distribution-Discrepancy Condensation and Manifold-Consistent Replay for Continual Face Forgery Detection

Tianshuo Zhang 1,2, Haoyuan Zhang 1,2, Siran Peng 2,1, Weisong Zhao 2,1, Xiangyu Zhu 1,2, Zhen Lei 1,2,3

0 citations

α

Published on arXiv

2604.12941

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Outperforms prior CFFD baselines under extremely small memory budgets while reducing identity leakage risk compared to selection-based replay methods

DDC-MCR

Novel technique introduced


Continual face forgery detection (CFFD) requires detectors to learn emerging forgery paradigms without forgetting previously seen manipulations. Existing CFFD methods commonly rely on replaying a small amount of past data to mitigate forgetting. Such replay is typically implemented either by storing a few historical samples or by synthesizing pseudo-forgeries from detector-dependent perturbations. Under strict memory budgets, the former cannot adequately cover diverse forgery cues and may expose facial identities, while the latter remains strongly tied to past decision boundaries. We argue that the core role of replay in CFFD is to reinstate the distributions of previous forgery tasks during subsequent training. To this end, we directly condense the discrepancy between real and fake distributions and leverage real faces from the current stage to perform distribution-level replay. Specifically, we introduce Distribution-Discrepancy Condensation (DDC), which models the real-to-fake discrepancy via a surrogate factorization in characteristic-function space and condenses it into a tiny bank of distribution discrepancy maps. We further propose Manifold-Consistent Replay (MCR), which synthesizes replay samples through variance-preserving composition of these maps with current-stage real faces, yielding samples that reflect previous-task forgery cues while remaining compatible with current real-face statistics. Operating under an extremely small memory budget and without directly storing raw historical face images, our framework consistently outperforms prior CFFD baselines and significantly mitigates catastrophic forgetting. Replay-level privacy analysis further suggests reduced identity leakage risk relative to selection-based replay.


Key Contributions

  • Distribution-Discrepancy Condensation (DDC) that models real-to-fake distribution gaps via characteristic functions and condenses them into compact distribution discrepancy maps
  • Manifold-Consistent Replay (MCR) that synthesizes replay samples by composing condensed discrepancy maps with current-stage real faces
  • Privacy-preserving continual learning framework that avoids storing raw historical face images while mitigating catastrophic forgetting

🛡️ Threat Analysis

Output Integrity Attack

Primary contribution is defending against AI-generated face forgeries (deepfakes) through continual detection — this is output integrity and AI-generated content detection. The paper builds a detector that identifies synthetic/manipulated faces across evolving forgery methods.


Details

Domains
visiongenerative
Model Types
gandiffusioncnn
Threat Tags
inference_time
Applications
deepfake detectionface forgery detection