Enhancing Adversarial Robustness of IoT Intrusion Detection via SHAP-Based Attribution Fingerprinting
Dilli Prasad Sharma 1, Liang Xue 1, Xiaowei Sun 1, Xiaodong Lin 2, Pulei Xiong 3
Published on arXiv
2511.06197
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
The SHAP attribution fingerprinting autoencoder significantly outperforms a state-of-the-art baseline in detecting adversarial attacks on an IoT IDS benchmark, with no requirement for labeled adversarial training data.
SHAP Attribution Fingerprinting
Novel technique introduced
The rapid proliferation of Internet of Things (IoT) devices has transformed numerous industries by enabling seamless connectivity and data-driven automation. However, this expansion has also exposed IoT networks to increasingly sophisticated security threats, including adversarial attacks targeting artificial intelligence (AI) and machine learning (ML)-based intrusion detection systems (IDS) to deliberately evade detection, induce misclassification, and systematically undermine the reliability and integrity of security defenses. To address these challenges, we propose a novel adversarial detection model that enhances the robustness of IoT IDS against adversarial attacks through SHapley Additive exPlanations (SHAP)-based fingerprinting. Using SHAP's DeepExplainer, we extract attribution fingerprints from network traffic features, enabling the IDS to reliably distinguish between clean and adversarially perturbed inputs. By capturing subtle attribution patterns, the model becomes more resilient to evasion attempts and adversarial manipulations. We evaluated the model on a standard IoT benchmark dataset, where it significantly outperformed a state-of-the-art method in detecting adversarial attacks. In addition to enhanced robustness, this approach improves model transparency and interpretability, thereby increasing trust in the IDS through explainable AI.
Key Contributions
- Novel unsupervised adversarial detection method using SHAP DeepExplainer attribution fingerprints and an autoencoder trained exclusively on clean samples to flag adversarial inputs via high reconstruction error — no labeled attack data required.
- Attribution fingerprinting pipeline that captures attack-specific feature importance rank shifts, revealing distinct behavioral patterns induced by different adversarial attacks.
- Demonstrated significant improvement over a state-of-the-art baseline in adversarial detection on an IoT benchmark dataset while also improving model interpretability.
🛡️ Threat Analysis
The paper defends against adversarial input manipulation at inference time — attackers craft adversarial perturbations to evade the ML-based IDS. The SHAP-based fingerprinting defense specifically detects adversarially perturbed inputs by identifying shifts in feature attribution patterns, a direct ML01 countermeasure.