attack arXiv Jan 15, 2026 · 11w ago
Frank Mollard, Marcus Becker, Florian Roehrbein · Infraserv · International School of Management +1 more
White-box adversarial attack using SHAP values to craft evasion perturbations, outperforming FGSM under gradient hiding
Input Manipulation Attack vision
The paper introduces a white-box attack on computer vision models using SHAP values. It demonstrates how adversarial evasion attacks can compromise the performance of deep learning models by reducing output confidence or inducing misclassifications. Such attacks are particularly insidious as they can deceive the perception of an algorithm while eluding human perception due to their imperceptibility to the human eye. The proposed attack leverages SHAP values to quantify the significance of individual inputs to the output at the inference stage. A comparison is drawn between the SHAP attack and the well-known Fast Gradient Sign Method. We find evidence that SHAP attacks are more robust in generating misclassifications particularly in gradient hiding scenarios.
cnn Infraserv · International School of Management · TU Chemnitz