defense 2026

Quadratic Upper Bound for Boosting Robustness

Euijin You , Hyang-Won Lee

0 citations · 45 references · arXiv

α

Published on arXiv

2601.13645

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Applying QUB loss to existing fast adversarial training methods yields significant robustness improvements while maintaining training efficiency, attributable to a smoothened perturbation loss landscape.

QUB (Quadratic Upper Bound)

Novel technique introduced


Fast adversarial training (FAT) aims to enhance the robustness of models against adversarial attacks with reduced training time, however, FAT often suffers from compromised robustness due to insufficient exploration of adversarial space. In this paper, we develop a loss function to mitigate the problem of degraded robustness under FAT. Specifically, we derive a quadratic upper bound (QUB) on the adversarial training (AT) loss function and propose to utilize the bound with existing FAT methods. Our experimental results show that applying QUB loss to the existing methods yields significant improvement of robustness. Furthermore, using various metrics, we demonstrate that this improvement is likely to result from the smoothened loss landscape of the resulting model.


Key Contributions

  • Derivation of a quadratic upper bound (QUB) on the adversarial training loss using the convexity of cross-entropy loss with respect to logits
  • A plug-in QUB loss that replaces the standard AT loss in existing fast adversarial training methods without substantially increasing training time
  • Empirical and metric-based analysis showing QUB improves robustness by smoothing the model's loss landscape with respect to perturbations

🛡️ Threat Analysis

Input Manipulation Attack

Proposes a defense against adversarial input manipulation attacks via a new adversarial training loss function (QUB); the paper directly addresses improving model robustness against FGSM/PGD-style adversarial examples at inference time.


Details

Domains
vision
Model Types
cnn
Threat Tags
white_boxinference_timeuntargeteddigital
Datasets
CIFAR-10CIFAR-100
Applications
image classification