Gradient Structure Estimation under Label-Only Oracles via Spectral Sensitivity
Jun Liu 1,2, Leo Yu Zhang 3, Fengpeng Li 1, Isao Echizen 4,2, Jiantao Zhou 1
Published on arXiv
2601.14300
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
DPAttack consistently surpasses SOTA hard-label black-box attacks in attack success rate and query efficiency across multiple benchmarks, achieving 0% detection rate against the Blacklight stateful defense.
Pattern-Driven Optimization (PDO) / DPAttack
Novel technique introduced
Hard-label black-box settings, where only top-1 predicted labels are observable, pose a fundamentally constrained yet practically important feedback model for understanding model behavior. A central challenge in this regime is whether meaningful gradient information can be recovered from such discrete responses. In this work, we develop a unified theoretical perspective showing that a wide range of existing sign-flipping hard-label attacks can be interpreted as implicitly approximating the sign of the true loss gradient. This observation reframes hard-label attacks from heuristic search procedures into instances of gradient sign recovery under extremely limited feedback. Motivated by this first-principles understanding, we propose a new attack framework that combines a zero-query frequency-domain initialization with a Pattern-Driven Optimization (PDO) strategy. We establish theoretical guarantees demonstrating that, under mild assumptions, our initialization achieves higher expected cosine similarity to the true gradient sign compared to random baselines, while the proposed PDO procedure attains substantially lower query complexity than existing structured search approaches. We empirically validate our framework through extensive experiments on CIFAR-10, ImageNet, and ObjectNet, covering standard and adversarially trained models, commercial APIs, and CLIP-based models. The results show that our method consistently surpasses SOTA hard-label attacks in both attack success rate and query efficiency, particularly in low-query regimes. Beyond image classification, our approach generalizes effectively to corrupted data, biomedical datasets, and dense prediction tasks. Notably, it also successfully circumvents Blacklight, a SOTA stateful defense, resulting in a $0\%$ detection rate. Our code will be released publicly soon at https://github.com/csjunjun/DPAttack.git.
Key Contributions
- Unified theoretical framework showing existing hard-label sign-flipping attacks implicitly approximate the true loss gradient sign, reframing them as gradient sign recovery problems.
- Pattern-Driven Optimization (PDO) attack framework with zero-query frequency-domain initialization that achieves higher cosine similarity to the true gradient sign and lower query complexity than prior structured search approaches.
- Empirical validation surpassing SOTA hard-label attacks across CIFAR-10, ImageNet, ObjectNet, and biomedical datasets, including 0% detection rate against the Blacklight stateful defense.
🛡️ Threat Analysis
Proposes a novel adversarial example attack (PDO/DPAttack) operating under hard-label black-box constraints, crafting inputs that cause misclassification at inference time by recovering gradient sign information from discrete label feedback — a canonical input manipulation attack.