defense arXiv Jan 21, 2026 · 10w ago
Liqin Wang, Qianyue Hu, Wei Lu et al. · Sun Yat-Sen University · State Key Laboratory of Mathematical Engineering and Advanced Computing
Adversarial perturbations that cascade-disrupt diffusion face-swapping pipelines by corrupting identity extraction and injection to prevent deepfakes
Input Manipulation Attack Output Integrity Attack visiongenerative
The rapid evolution of diffusion models has democratized face swapping but also raises concerns about privacy and identity security. Existing proactive defenses, often adapted from image editing attacks, prove ineffective in this context. We attribute this failure to an oversight of the structural resilience and the unique static conditional guidance mechanism inherent in face swapping systems. To address this, we propose VoidFace, a systemic defense method that views face swapping as a coupled identity pathway. By injecting perturbations at critical bottlenecks, VoidFace induces cascading disruption throughout the pipeline. Specifically, we first introduce localization disruption and identity erasure to degrade physical regression and semantic embeddings, thereby impairing the accurate modeling of the source face. We then intervene in the generative domain by decoupling attention mechanisms to sever identity injection, and corrupting intermediate diffusion features to prevent the reconstruction of source identity. To ensure visual imperceptibility, we perform adversarial search in the latent manifold, guided by a perceptual adaptive strategy to balance attack potency with image quality. Extensive experiments show that VoidFace outperforms existing defenses across various diffusion-based swapping models, while producing adversarial faces with superior visual quality.
diffusion Sun Yat-Sen University · State Key Laboratory of Mathematical Engineering and Advanced Computing
defense arXiv Jan 29, 2026 · 9w ago
Lingxiao Chen, Liqin Wang, Wei Lu et al. · Sun Yat-Sen University · State Key Laboratory of Mathematical Engineering and Advanced Computing
Fingerprints diffusion models via denoising trajectory manifolds to verify copyright in black-box API settings without model modification
Model Theft visiongenerative
The exceptional performance of diffusion models establishes them as high-value intellectual property but exposes them to unauthorized replication. Existing protection methods either modify the model to embed watermarks, which impairs performance, or extract model fingerprints by manipulating the denoising process, rendering them incompatible with black-box APIs. In this paper, we propose TrajPrint, a completely lossless and training-free framework that verifies model copyright by extracting unique manifold fingerprints formed during deterministic generation. Specifically, we first utilize a watermarked image as an anchor and exactly trace the path back to its trajectory origin, effectively locking the model fingerprint mapped by this path. Subsequently, we implement a joint optimization strategy that employs dual-end anchoring to synthesize a specific fingerprint noise, which strictly adheres to the target manifold for robust watermark recovery. As input, it enables the protected target model to recover the watermarked image, while failing on non-target models. Finally, we achieved verification via atomic inference and statistical hypothesis testing. Extensive experiments demonstrate that TrajPrint achieves lossless verification in black-box API scenarios with superior robustness against model modifications.
diffusion Sun Yat-Sen University · State Key Laboratory of Mathematical Engineering and Advanced Computing