defense 2025

MANI-Pure: Magnitude-Adaptive Noise Injection for Adversarial Purification

Xiaoyi Huang 1, Junwei Wu 2, Kejia Zhang 1, Carl Yang 2, Zhiming Luo 1

0 citations · 39 references · arXiv

α

Published on arXiv

2509.25082

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Achieves top-1 robust accuracy on RobustBench leaderboard, boosting robust accuracy by 2.15% while narrowing the clean accuracy gap to within 0.59% of the undefended classifier.

MANI-Pure

Novel technique introduced


Adversarial purification with diffusion models has emerged as a promising defense strategy, but existing methods typically rely on uniform noise injection, which indiscriminately perturbs all frequencies, corrupting semantic structures and undermining robustness. Our empirical study reveals that adversarial perturbations are not uniformly distributed: they are predominantly concentrated in high-frequency regions, with heterogeneous magnitude intensity patterns that vary across frequencies and attack types. Motivated by this observation, we introduce MANI-Pure, a magnitude-adaptive purification framework that leverages the magnitude spectrum of inputs to guide the purification process. Instead of injecting homogeneous noise, MANI-Pure adaptively applies heterogeneous, frequency-targeted noise, effectively suppressing adversarial perturbations in fragile high-frequency, low-magnitude bands while preserving semantically critical low-frequency content. Extensive experiments on CIFAR-10 and ImageNet-1K validate the effectiveness of MANI-Pure. It narrows the clean accuracy gap to within 0.59 of the original classifier, while boosting robust accuracy by 2.15, and achieves the top-1 robust accuracy on the RobustBench leaderboard, surpassing the previous state-of-the-art method.


Key Contributions

  • Empirical analysis revealing adversarial perturbations concentrate in high-frequency, low-magnitude spectral bands across multiple attack types
  • MANI-Pure: a magnitude-adaptive purification framework that injects heterogeneous, frequency-targeted noise guided by the input's magnitude spectrum instead of uniform noise
  • State-of-the-art top-1 robust accuracy on RobustBench leaderboard for CIFAR-10 and ImageNet-1K, boosting robust accuracy by 2.15% while keeping clean accuracy within 0.59% of the original classifier

🛡️ Threat Analysis

Input Manipulation Attack

Proposes a defense against adversarial examples at inference time — MANI-Pure purifies adversarially perturbed inputs using magnitude-adaptive, frequency-targeted diffusion noise injection before classification, directly countering input manipulation attacks.


Details

Domains
vision
Model Types
diffusioncnn
Threat Tags
white_boxblack_boxinference_timedigitaluntargeted
Datasets
CIFAR-10ImageNet-1KRobustBench
Applications
image classification