attack 2025

AIM: Additional Image Guided Generation of Transferable Adversarial Attacks

Teng Li , Xingjun Ma , Yu-Gang Jiang

5 citations · 1 influential · 37 references · AAAI

α

Published on arXiv

2501.01106

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

AIM achieves superior targeted transferable attack success rates compared to state-of-the-art generative methods while maintaining competitive untargeted attack performance.

AIM / Semantic Injection Module (SIM)

Novel technique introduced


Transferable adversarial examples highlight the vulnerability of deep neural networks (DNNs) to imperceptible perturbations across various real-world applications. While there have been notable advancements in untargeted transferable attacks, targeted transferable attacks remain a significant challenge. In this work, we focus on generative approaches for targeted transferable attacks. Current generative attacks focus on reducing overfitting to surrogate models and the source data domain, but they often overlook the importance of enhancing transferability through additional semantics. To address this issue, we introduce a novel plug-and-play module into the general generator architecture to enhance adversarial transferability. Specifically, we propose a \emph{Semantic Injection Module} (SIM) that utilizes the semantics contained in an additional guiding image to improve transferability. The guiding image provides a simple yet effective method to incorporate target semantics from the target class to create targeted and highly transferable attacks. Additionally, we propose new loss formulations that can integrate the semantic injection module more effectively for both targeted and untargeted attacks. We conduct comprehensive experiments under both targeted and untargeted attack settings to demonstrate the efficacy of our proposed approach.


Key Contributions

  • Proposes the Semantic Injection Module (SIM), a lightweight plug-and-play component that conditions an adversarial generator on a guiding image to inject target-class semantics and improve transferability.
  • Introduces new loss formulations (logit contrastive loss and mid-layer similarity loss) tailored for both targeted and untargeted generative attacks using SIM.
  • Demonstrates superior targeted transferability over prior generative attack methods and competitive untargeted performance via comprehensive experiments.

🛡️ Threat Analysis

Input Manipulation Attack

Proposes a generative attack that crafts imperceptible adversarial perturbations causing misclassification at inference time, with the explicit goal of transferability to unseen black-box target models — core adversarial example / input manipulation attack.


Details

Domains
vision
Model Types
cnntransformer
Threat Tags
black_boxinference_timetargeteduntargeteddigital
Datasets
ImageNet
Applications
image classification