attack 2025

Everywhere Attack: Attacking Locally and Globally to Boost Targeted Transferability

Hui Zeng 1,2, Sanshuai Cui 3, Biwei Chen 4, Anjie Peng 1

5 citations

α

Published on arXiv

2501.00707

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Improves targeted transferability of the Logit attack by 28.8%–300% on ImageNet across multiple black-box victim models.

Everywhere Attack

Novel technique introduced


Adversarial examples' (AE) transferability refers to the phenomenon that AEs crafted with one surrogate model can also fool other models. Notwithstanding remarkable progress in untargeted transferability, its targeted counterpart remains challenging. This paper proposes an everywhere scheme to boost targeted transferability. Our idea is to attack a victim image both globally and locally. We aim to optimize 'an army of targets' in every local image region instead of the previous works that optimize a high-confidence target in the image. Specifically, we split a victim image into non-overlap blocks and jointly mount a targeted attack on each block. Such a strategy mitigates transfer failures caused by attention inconsistency between surrogate and victim models and thus results in stronger transferability. Our approach is method-agnostic, which means it can be easily combined with existing transferable attacks for even higher transferability. Extensive experiments on ImageNet demonstrate that the proposed approach universally improves the state-of-the-art targeted attacks by a clear margin, e.g., the transferability of the widely adopted Logit attack can be improved by 28.8%-300%.We also evaluate the crafted AEs on a real-world platform: Google Cloud Vision. Results further support the superiority of the proposed method.


Key Contributions

  • Identifies attention inconsistency between surrogate and victim models as a primary cause of targeted transfer failure.
  • Proposes the 'Everywhere Attack' which simultaneously mounts targeted attacks on every non-overlapping local block of a victim image, covering diverse attention regions of victim models.
  • Method-agnostic scheme that universally improves state-of-the-art targeted attacks (e.g., Logit attack by 28.8%–300%) and is validated on both ImageNet and the Google Cloud Vision API.

🛡️ Threat Analysis

Input Manipulation Attack

Proposes gradient-based adversarial perturbations crafted on surrogate models to cause targeted misclassification on unseen victim models at inference time — core adversarial example / evasion attack methodology.


Details

Domains
vision
Model Types
cnntransformer
Threat Tags
black_boxinference_timetargeteddigital
Datasets
ImageNetGoogle Cloud Vision
Applications
image classification