attack 2026

BadSNN: Backdoor Attacks on Spiking Neural Networks via Adversarial Spiking Neuron

Abdullah Arafat Miah , Kevin Vu , Yu Bi

0 citations · 28 references · arXiv (Cornell University)

α

Published on arXiv

2602.07200

Model Poisoning

OWASP ML Top 10 — ML10

Key Finding

BadSNN achieves superior backdoor attack success rate on SNNs compared to state-of-the-art data-poisoning-based backdoor attacks while remaining robust against common backdoor mitigation defenses.

BadSNN

Novel technique introduced


Spiking Neural Networks (SNNs) are energy-efficient counterparts of Deep Neural Networks (DNNs) with high biological plausibility, as information is transmitted through temporal spiking patterns. The core element of an SNN is the spiking neuron, which converts input data into spikes following the Leaky Integrate-and-Fire (LIF) neuron model. This model includes several important hyperparameters, such as the membrane potential threshold and membrane time constant. Both the DNNs and SNNs have proven to be exploitable by backdoor attacks, where an adversary can poison the training dataset with malicious triggers and force the model to behave in an attacker-defined manner. Yet, how an adversary can exploit the unique characteristics of SNNs for backdoor attacks remains underexplored. In this paper, we propose \textit{BadSNN}, a novel backdoor attack on spiking neural networks that exploits hyperparameter variations of spiking neurons to inject backdoor behavior into the model. We further propose a trigger optimization process to achieve better attack performance while making trigger patterns less perceptible. \textit{BadSNN} demonstrates superior attack performance on various datasets and architectures, as well as compared with state-of-the-art data poisoning-based backdoor attacks and robustness against common backdoor mitigation techniques. Codes can be found at https://github.com/SiSL-URI/BadSNN.


Key Contributions

  • First backdoor attack exploiting SNN-specific hyperparameter variations (LIF neuron membrane potential threshold and time constant) as the attack vector
  • Trigger optimization process that improves attack success rate while minimizing trigger perceptibility
  • Demonstrated superior attack performance and robustness against common backdoor mitigation techniques across multiple datasets and SNN architectures

🛡️ Threat Analysis

Model Poisoning

BadSNN is a backdoor injection attack: the adversary poisons training data with malicious triggers and exploits SNN-specific hyperparameters (membrane potential threshold, membrane time constant) to embed hidden targeted behavior that activates only on trigger inputs while the model behaves normally otherwise.


Details

Domains
vision
Model Types
cnn
Threat Tags
training_timetargeteddigital
Applications
image classificationspiking neural networks