The Power of Decaying Steps: Enhancing Attack Stability and Transferability for Sign-based Optimizers
Wei Tao 1,2, Yang Dai 1, Jincai Huang 1, Qing Tao 3
Published on arXiv
2602.19096
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
MDCS-MI achieves a provably optimal O(1/√T) convergence rate and significantly outperforms MI-FGSM and other sign-based baselines in transfer attack success rate and stability across image classification and cross-modal retrieval benchmarks.
MDCS-MI
Novel technique introduced
Crafting adversarial examples can be formulated as an optimization problem. While sign-based optimizers such as I-FGSM and MI-FGSM have become the de facto standard for the induced optimization problems, there still exist several unsolved problems in theoretical grounding and practical reliability especially in non-convergence and instability, which inevitably influences their transferability. Contrary to the expectation, we observe that the attack success rate may degrade sharply when more number of iterations are conducted. In this paper, we address these issues from an optimization perspective. By reformulating the sign-based optimizer as a specific coordinate-wise gradient descent, we argue that one cause for non-convergence and instability is their non-decaying step-size scheduling. Based upon this viewpoint, we propose a series of new attack algorithms that enforce Monotonically Decreasing Coordinate-wise Step-sizes (MDCS) within sign-based optimizers. Typically, we further provide theoretical guarantees proving that MDCS-MI attains an optimal convergence rate of $O(1/\sqrt{T})$, where $T$ is the number of iterations. Extensive experiments on image classification and cross-modal retrieval tasks demonstrate that our approach not only significantly improves transferability but also enhances attack stability compared to state-of-the-art sign-based methods.
Key Contributions
- Reformulates sign-based adversarial optimizers (I-FGSM, MI-FGSM) as coordinate-wise gradient descent, identifying non-decaying step-size as a root cause of instability and non-convergence
- Proposes MDCS (Monotonically Decreasing Coordinate-wise Step-sizes) attack family and proves MDCS-MI achieves optimal O(1/√T) convergence rate
- Demonstrates improved adversarial transferability and attack stability over state-of-the-art sign-based methods on image classification and cross-modal retrieval tasks
🛡️ Threat Analysis
Directly proposes new gradient-based adversarial example generation algorithms (MDCS-MI, MDCS variants) that improve upon I-FGSM and MI-FGSM to cause misclassification at inference time, with explicit focus on black-box transferability.