defense 2026

Robust Spiking Neural Networks Against Adversarial Attacks

Shuai Wang 1, Malu Zhang 1,2, Yulin Jiang 1, Dehao Zhang 1, Ammar Belatreche 3, Yu Liang 1, Yimeng Shan 1, Zijian Zhou 1, Yang Yang 1, Haizhou Li 2,4

0 citations · 81 references · arXiv (Cornell University)

α

Published on arXiv

2602.20548

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

TGO achieves state-of-the-art adversarial robustness for directly trained SNNs across multiple attack scenarios with no additional computational overhead at inference time

Threshold Guarding Optimization (TGO)

Novel technique introduced


Spiking Neural Networks (SNNs) represent a promising paradigm for energy-efficient neuromorphic computing due to their bio-plausible and spike-driven characteristics. However, the robustness of SNNs in complex adversarial environments remains significantly constrained. In this study, we theoretically demonstrate that those threshold-neighboring spiking neurons are the key factors limiting the robustness of directly trained SNNs. We find that these neurons set the upper limits for the maximum potential strength of adversarial attacks and are prone to state-flipping under minor disturbances. To address this challenge, we propose a Threshold Guarding Optimization (TGO) method, which comprises two key aspects. First, we incorporate additional constraints into the loss function to move neurons' membrane potentials away from their thresholds. It increases SNNs' gradient sparsity, thereby reducing the theoretical upper bound of adversarial attacks. Second, we introduce noisy spiking neurons to transition the neuronal firing mechanism from deterministic to probabilistic, decreasing their state-flipping probability due to minor disturbances. Extensive experiments conducted in standard adversarial scenarios prove that our method significantly enhances the robustness of directly trained SNNs. These findings pave the way for advancing more reliable and secure neuromorphic computing in real-world applications.


Key Contributions

  • Theoretical analysis showing threshold-neighboring spiking neurons are the primary robustness bottleneck in directly trained SNNs, defining upper limits on adversarial attack strength
  • Threshold Guarding Optimization (TGO): loss constraints that push membrane potentials away from firing thresholds, increasing gradient sparsity and reducing adversarial attack potential
  • Introduction of noisy spiking neurons that convert deterministic firing to probabilistic, reducing state-flipping probability under minor disturbances with zero inference-time overhead

🛡️ Threat Analysis

Input Manipulation Attack

Proposes a defense (Threshold Guarding Optimization) specifically against adversarial input perturbations that cause misclassification in SNNs at inference time — directly addresses gradient-based evasion attacks like FGSM and PGD.


Details

Domains
vision
Model Types
cnn
Threat Tags
white_boxinference_timeuntargeteddigital
Datasets
CIFAR-10
Applications
image classificationneuromorphic computing