StegaFFD: Privacy-Preserving Face Forgery Detection via Fine-Grained Steganographic Domain Lifting
Guoqing Ma 1,2, Xun Lin 3, Hui Ma 4, Ajian Liu 1,4, Yizhong Liu 3, Wenzhong Tang 3, Shan Yu 1,2, Chenqi Kong 5, Yi Yu 5
Published on arXiv
2603.02886
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
StegaFFD achieves strong imperceptibility and better preserves face forgery detection accuracy compared to anonymization, encryption, and distortion-based privacy protection methods across seven FFD datasets.
StegaFFD
Novel technique introduced
Most existing Face Forgery Detection (FFD) models assume access to raw face images. In practice, under a client-server framework, private facial data may be intercepted during transmission or leaked by untrusted servers. Previous privacy protection approaches, such as anonymization, encryption, or distortion, partly mitigate leakage but often introduce severe semantic distortion, making images appear obviously protected. This alerts attackers, provoking more aggressive strategies and turning the process into a cat-and-mouse game. Moreover, these methods heavily manipulate image contents, introducing degradation or artifacts that may confuse FFD models, which rely on extremely subtle forgery traces. Inspired by advances in image steganography, which enable high-fidelity hiding and recovery, we propose a Stega}nography-based Face Forgery Detection framework (StegaFFD) to protect privacy without raising suspicion. StegaFFD hides facial images within natural cover images and directly conducts forgery detection in the steganographic domain. However, the hidden forgery-specific features are extremely subtle and interfered with by cover semantics, posing significant challenges. To address this, we propose Low-Frequency-Aware Decomposition (LFAD) and Spatial-Frequency Differential Attention (SFDA), which suppress interference from low-frequency cover semantics and enhance hidden facial feature perception. Furthermore, we introduce Steganographic Domain Alignment (SDA) to align the representations of hidden faces with those of their raw counterparts, enhancing the model's ability to perceive subtle facial cues in the steganographic domain. Extensive experiments on seven FFD datasets demonstrate that StegaFFD achieves strong imperceptibility, avoids raising attackers' suspicion, and better preserves FFD accuracy compared to existing facial privacy protection methods.
Key Contributions
- StegaFFD framework that hides facial images within natural cover images using steganography and performs face forgery detection directly in the steganographic domain, avoiding detection by adversaries who intercept transmissions
- Low-Frequency-Aware Decomposition (LFAD) and Spatial-Frequency Differential Attention (SFDA) to suppress cover image semantic interference and enhance perception of hidden face forgery traces
- Steganographic Domain Alignment (SDA) to align hidden face representations with raw face representations, improving the model's ability to detect subtle forgery cues in the steganographic domain
🛡️ Threat Analysis
The paper's primary ML contribution is face forgery detection — detecting AI-manipulated/generated facial content (deepfakes). It proposes novel detection architectures (LFAD, SFDA, SDA) that operate in the steganographic domain, directly addressing AI-generated content detection. The privacy-preserving steganographic hiding is the delivery mechanism, but the core ML security contribution is deepfake/face forgery detection, which falls squarely under ML09 output integrity and AI-generated content detection.