attack 2025

Investigating Adversarial Robustness against Preprocessing used in Blackbox Face Recognition

Roland Croft , Brian Du , Darcy Joseph , Sharath Kumar

0 citations · 45 references · International Conference on Di...

α

Published on arXiv

2510.17169

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Face detection model choice alone degrades adversarial attack success rates by up to 78%, and the proposed preprocessing-invariant augmentation recovers up to 27% transferability across blackbox FR systems.

Preprocessing-Invariant Input Transformation

Novel technique introduced


Face Recognition (FR) models have been shown to be vulnerable to adversarial examples that subtly alter benign facial images, exposing blind spots in these systems, as well as protecting user privacy. End-to-end FR systems first obtain preprocessed faces from diverse facial imagery prior to computing the similarity of the deep feature embeddings. Whilst face preprocessing is a critical component of FR systems, and hence adversarial attacks against them, we observe that this preprocessing is often overlooked in blackbox settings. Our study seeks to investigate the transferability of several out-of-the-box state-of-the-art adversarial attacks against FR when applied against different preprocessing techniques used in a blackbox setting. We observe that the choice of face detection model can degrade the attack success rate by up to 78%, whereas choice of interpolation method during downsampling has relatively minimal impacts. Furthermore, we find that the requirement for facial preprocessing even degrades attack strength in a whitebox setting, due to the unintended interaction of produced noise vectors against face detection models. Based on these findings, we propose a preprocessing-invariant method using input transformations that improves the transferability of the studied attacks by up to 27%. Our findings highlight the importance of preprocessing in FR systems, and the need for its consideration towards improving the adversarial generalisation of facial adversarial examples.


Key Contributions

  • Empirical analysis showing face detection model choice can degrade adversarial attack success rate by up to 78% in blackbox FR settings, while interpolation method has minimal impact
  • Finding that face preprocessing degrades adversarial strength even in whitebox settings due to unintended noise-detector interactions
  • Preprocessing-invariant attack augmentation using input transformations that improves adversarial transferability by up to 27% across different preprocessing pipelines

🛡️ Threat Analysis

Input Manipulation Attack

Paper studies evasion/adversarial example attacks against face recognition systems at inference time, analyzes how preprocessing pipelines (face detection models, interpolation) degrade adversarial perturbation transferability in blackbox settings, and proposes a preprocessing-invariant input transformation method to improve adversarial transferability by up to 27%.


Details

Domains
vision
Model Types
cnn
Threat Tags
black_boxwhite_boxinference_timetargeteddigital
Applications
face recognitionfacial privacy protection