attack 2026

The Power of Decaying Steps: Enhancing Attack Stability and Transferability for Sign-based Optimizers

Wei Tao 1,2, Yang Dai 1, Jincai Huang 1, Qing Tao 3

0 citations · 51 references · arXiv (Cornell University)

α

Published on arXiv

2602.19096

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

MDCS-MI achieves a provably optimal O(1/√T) convergence rate and significantly outperforms MI-FGSM and other sign-based baselines in transfer attack success rate and stability across image classification and cross-modal retrieval benchmarks.

MDCS-MI

Novel technique introduced


Crafting adversarial examples can be formulated as an optimization problem. While sign-based optimizers such as I-FGSM and MI-FGSM have become the de facto standard for the induced optimization problems, there still exist several unsolved problems in theoretical grounding and practical reliability especially in non-convergence and instability, which inevitably influences their transferability. Contrary to the expectation, we observe that the attack success rate may degrade sharply when more number of iterations are conducted. In this paper, we address these issues from an optimization perspective. By reformulating the sign-based optimizer as a specific coordinate-wise gradient descent, we argue that one cause for non-convergence and instability is their non-decaying step-size scheduling. Based upon this viewpoint, we propose a series of new attack algorithms that enforce Monotonically Decreasing Coordinate-wise Step-sizes (MDCS) within sign-based optimizers. Typically, we further provide theoretical guarantees proving that MDCS-MI attains an optimal convergence rate of $O(1/\sqrt{T})$, where $T$ is the number of iterations. Extensive experiments on image classification and cross-modal retrieval tasks demonstrate that our approach not only significantly improves transferability but also enhances attack stability compared to state-of-the-art sign-based methods.


Key Contributions

  • Reformulates sign-based adversarial optimizers (I-FGSM, MI-FGSM) as coordinate-wise gradient descent, identifying non-decaying step-size as a root cause of instability and non-convergence
  • Proposes MDCS (Monotonically Decreasing Coordinate-wise Step-sizes) attack family and proves MDCS-MI achieves optimal O(1/√T) convergence rate
  • Demonstrates improved adversarial transferability and attack stability over state-of-the-art sign-based methods on image classification and cross-modal retrieval tasks

🛡️ Threat Analysis

Input Manipulation Attack

Directly proposes new gradient-based adversarial example generation algorithms (MDCS-MI, MDCS variants) that improve upon I-FGSM and MI-FGSM to cause misclassification at inference time, with explicit focus on black-box transferability.


Details

Domains
visionmultimodal
Model Types
cnntransformer
Threat Tags
white_boxblack_boxinference_timeuntargeteddigital
Datasets
ImageNet
Applications
image classificationcross-modal retrieval