attack 2025

ViT-EnsembleAttack: Augmenting Ensemble Models for Stronger Adversarial Transferability in Vision Transformers

Hanwen Cao , Haobo Lu , Xiaosen Wang , Kun He

0 citations

α

Published on arXiv

2508.12384

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

ViT-EnsembleAttack significantly outperforms existing ensemble-based transfer attack methods on ViT architectures across extensive experiments.

ViT-EnsembleAttack

Novel technique introduced


Ensemble-based attacks have been proven to be effective in enhancing adversarial transferability by aggregating the outputs of models with various architectures. However, existing research primarily focuses on refining ensemble weights or optimizing the ensemble path, overlooking the exploration of ensemble models to enhance the transferability of adversarial attacks. To address this gap, we propose applying adversarial augmentation to the surrogate models, aiming to boost overall generalization of ensemble models and reduce the risk of adversarial overfitting. Meanwhile, observing that ensemble Vision Transformers (ViTs) gain less attention, we propose ViT-EnsembleAttack based on the idea of model adversarial augmentation, the first ensemble-based attack method tailored for ViTs to the best of our knowledge. Our approach generates augmented models for each surrogate ViT using three strategies: Multi-head dropping, Attention score scaling, and MLP feature mixing, with the associated parameters optimized by Bayesian optimization. These adversarially augmented models are ensembled to generate adversarial examples. Furthermore, we introduce Automatic Reweighting and Step Size Enlargement modules to boost transferability. Extensive experiments demonstrate that ViT-EnsembleAttack significantly enhances the adversarial transferability of ensemble-based attacks on ViTs, outperforming existing methods by a substantial margin. Code is available at https://github.com/Trustworthy-AI-Group/TransferAttack.


Key Contributions

  • Three ViT-specific model augmentation strategies (Multi-head dropping, Attention score scaling, MLP feature mixing) with Bayesian-optimized parameters to diversify surrogate ensembles and reduce adversarial overfitting
  • Automatic Reweighting and Step Size Enlargement modules to further improve adversarial transferability across ensemble ViTs
  • First ensemble-based transfer attack method specifically tailored for Vision Transformers, outperforming existing methods by a substantial margin

🛡️ Threat Analysis

Input Manipulation Attack

Core contribution is crafting adversarial examples that transfer across ViT architectures via gradient-based optimization on augmented surrogate ensembles — a direct inference-time input manipulation attack.


Details

Domains
vision
Model Types
transformer
Threat Tags
white_boxblack_boxinference_timedigital
Applications
image classification