attack 2025

Real-World Transferable Adversarial Attack on Face-Recognition Systems

Andrey Kaznacheev 1, Matvey Mikhalchuk 2, Andrey Kuznetsov 3,2, Aleksandr Petiushko 4, Anton Razzhigaev 2,1

0 citations · 18 references · arXiv

α

Published on arXiv

2509.23198

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

GaP achieves high attack success rate in real-world physical tests against ArcFace with ~10,000 black-box queries and transfers successfully to an unseen FaceNet model

GaP (Gaussian Patch)

Novel technique introduced


Adversarial attacks on face recognition (FR) systems pose a significant security threat, yet most are confined to the digital domain or require white-box access. We introduce GaP (Gaussian Patch), a novel method to generate a universal, physically transferable adversarial patch under a strict black-box setting. Our approach uses a query-efficient, zero-order greedy algorithm to iteratively construct a symmetric, grayscale pattern for the forehead. The patch is optimized by successively adding Gaussian blobs, guided only by the cosine similarity scores from a surrogate FR model to maximally degrade identity recognition. We demonstrate that with approximately 10,000 queries to a black-box ArcFace model, the resulting GaP achieves a high attack success rate in both digital and real-world physical tests. Critically, the attack shows strong transferability, successfully deceiving an entirely unseen FaceNet model. Our work highlights a practical and severe vulnerability, proving that robust, transferable attacks can be crafted with limited knowledge of the target system.


Key Contributions

  • GaP: a query-efficient zero-order greedy algorithm that iteratively builds a symmetric grayscale forehead patch by successively adding Gaussian blobs guided by cosine similarity scores from a black-box surrogate FR model
  • Demonstrates high physical attack success rate with ~10,000 queries to ArcFace in both digital and print-and-wear real-world trials
  • Shows strong transferability of the generated patch to an entirely unseen FaceNet model despite optimization on ArcFace only

🛡️ Threat Analysis

Input Manipulation Attack

GaP generates physical adversarial patches worn on the forehead that degrade face recognition at inference time — a direct adversarial patch attack optimized via zero-order black-box queries, effective in both digital and real-world physical settings with transferability to unseen models.


Details

Domains
vision
Model Types
cnn
Threat Tags
black_boxinference_timeuntargetedphysicaldigital
Datasets
CelebA
Applications
face recognitionbiometric authentication