defense 2025

Filtered-ViT: A Robust Defense Against Multiple Adversarial Patch Attacks

Aja Khanal , Ahmed Faid , Apurva Narayan

0 citations · 19 references · Symposium Series

α

Published on arXiv

2511.07755

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Achieves 79.8% clean accuracy and 46.3% robust accuracy under four simultaneous 1% adversarial patches on ImageNet, outperforming existing defenses.

Filtered-ViT / SMART-VMF

Novel technique introduced


Deep learning vision systems are increasingly deployed in safety-critical domains such as healthcare, yet they remain vulnerable to small adversarial patches that can trigger misclassifications. Most existing defenses assume a single patch and fail when multiple localized disruptions occur, the type of scenario adversaries and real-world artifacts often exploit. We propose Filtered-ViT, a new vision transformer architecture that integrates SMART Vector Median Filtering (SMART-VMF), a spatially adaptive, multi-scale, robustness-aware mechanism that enables selective suppression of corrupted regions while preserving semantic detail. On ImageNet with LaVAN multi-patch attacks, Filtered-ViT achieves 79.8% clean accuracy and 46.3% robust accuracy under four simultaneous 1\% patches, outperforming existing defenses. Beyond synthetic benchmarks, a real-world case study on radiographic medical imagery shows that Filtered-ViT mitigates natural artifacts such as occlusions and scanner noise without degrading diagnostic content. This establishes Filtered-ViT as the first transformer to demonstrate unified robustness against both adversarial and naturally occurring patch-like disruptions, charting a path toward reliable vision systems in truly high-stakes environments.


Key Contributions

  • Filtered-ViT: a Vision Transformer architecture integrating SMART Vector Median Filtering (SMART-VMF) to selectively suppress adversarial patch regions while preserving semantic content
  • First demonstrated defense against multiple simultaneous adversarial patches (up to four 1% patches under LaVAN attack) in a unified transformer framework
  • Real-world case study on radiographic medical imagery showing the defense also mitigates natural artifacts like occlusions and scanner noise

🛡️ Threat Analysis

Input Manipulation Attack

Directly defends against adversarial patch attacks — localized, physically realizable perturbations that cause misclassification at inference time. SMART-VMF is designed to suppress these corrupted regions, which are a canonical form of input manipulation attack.


Details

Domains
vision
Model Types
transformer
Threat Tags
inference_timedigitaluntargeted
Datasets
ImageNet
Applications
image classificationmedical imaging