defense 2025

DevFD: Developmental Face Forgery Detection by Learning Shared and Orthogonal LoRA Subspaces

Tianshuo Zhang 1,2, Li Gao 3, Siran Peng 1,2, Xiangyu Zhu 1,2, Zhen Lei 1,2,2,4

0 citations · 70 references · arXiv

α

Published on arXiv

2509.19230

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Achieves state-of-the-art average accuracy and lowest average forgetting rate under both dataset-incremental and manipulation-type-incremental continual learning protocols.

DevFD

Novel technique introduced


The rise of realistic digital face generation and manipulation poses significant social risks. The primary challenge lies in the rapid and diverse evolution of generation techniques, which often outstrip the detection capabilities of existing models. To defend against the ever-evolving new types of forgery, we need to enable our model to quickly adapt to new domains with limited computation and data while avoiding forgetting previously learned forgery types. In this work, we posit that genuine facial samples are abundant and relatively stable in acquisition methods, while forgery faces continuously evolve with the iteration of manipulation techniques. Given the practical infeasibility of exhaustively collecting all forgery variants, we frame face forgery detection as a continual learning problem and allow the model to develop as new forgery types emerge. Specifically, we employ a Developmental Mixture of Experts (MoE) architecture that uses LoRA models as its individual experts. These experts are organized into two groups: a Real-LoRA to learn and refine knowledge of real faces, and multiple Fake-LoRAs to capture incremental information from different forgery types. To prevent catastrophic forgetting, we ensure that the learning direction of Fake-LoRAs is orthogonal to the established subspace. Moreover, we integrate orthogonal gradients into the orthogonal loss of Fake-LoRAs, preventing gradient interference throughout the training process of each task. Experimental results under both the datasets and manipulation types incremental protocols demonstrate the effectiveness of our method.


Key Contributions

  • Developmental Mixture of Experts (MoE) architecture using LoRA experts split into a shared Real-LoRA for genuine faces and orthogonal Fake-LoRAs for incrementally emerging forgery types
  • Integrated orthogonal loss that combines subspace orthogonal loss with orthogonal gradient projection to prevent catastrophic forgetting throughout all training stages
  • Label-guided localized balancing strategy that softly constrains expert routing to separate real-face modeling from incremental forgery-type adaptation

🛡️ Threat Analysis

Output Integrity Attack

Proposes a novel face forgery detection system that classifies images as real or AI-generated/manipulated — directly addressing AI-generated content detection and output integrity. The primary contribution is the DevFD architecture for detecting deepfakes across evolving forgery types.


Details

Domains
vision
Model Types
transformer
Threat Tags
inference_time
Datasets
FaceForensics++ (FF++)CelebDF-v2 (CDF2)
Applications
face forgery detectiondeepfake detection