defense 2025

Tuning for Two Adversaries: Enhancing the Robustness Against Transfer and Query-Based Attacks using Hyperparameter Tuning

Pascal Zimmer , Ghassan Karame

0 citations · 40 references · arXiv

α

Published on arXiv

2511.13654

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Distributed models achieve the best joint robustness tradeoff via hyperparameter tuning, simultaneously mitigating transfer-based attacks by up to 64% and query-based attacks by up to 28%

Hyperparameter Tuning for Dual Adversarial Robustness

Novel technique introduced


In this paper, we present the first detailed analysis of how training hyperparameters -- such as learning rate, weight decay, momentum, and batch size -- influence robustness against both transfer-based and query-based attacks. Supported by theory and experiments, our study spans a variety of practical deployment settings, including centralized training, ensemble learning, and distributed training. We uncover a striking dichotomy: for transfer-based attacks, decreasing the learning rate significantly enhances robustness by up to $64\%$. In contrast, for query-based attacks, increasing the learning rate consistently leads to improved robustness by up to $28\%$ across various settings and data distributions. Leveraging these findings, we explore -- for the first time -- the training hyperparameter space to jointly enhance robustness against both transfer-based and query-based attacks. Our results reveal that distributed models benefit the most from hyperparameter tuning, achieving a remarkable tradeoff by simultaneously mitigating both attack types more effectively than other training setups.


Key Contributions

  • First systematic analysis of how training hyperparameters (learning rate, weight decay, momentum, batch size) affect robustness against both transfer-based and query-based attacks
  • Discovery of a dichotomy: lower learning rate improves robustness against transfer attacks (up to 64%) while higher learning rate improves robustness against query attacks (up to 28%)
  • Multi-objective hyperparameter optimization (via NSGA-II) to jointly enhance robustness against both attack types, with distributed models benefiting most

🛡️ Threat Analysis

Input Manipulation Attack

Directly addresses inference-time evasion attacks — both transfer-based (surrogate black-box) and query-based (decision/score black-box) adversarial examples — proposing hyperparameter tuning as a defense mechanism that improves robustness by up to 64% and 28% respectively.


Details

Domains
visionfederated-learning
Model Types
cnnfederated
Threat Tags
black_boxinference_timeuntargeted
Datasets
CIFAR-10CIFAR-100
Applications
image classification