defense 2025

If You Want to Be Robust, Be Wary of Initialization

Sofiane Ennadir 1, Johannes F. Lutzeyer 2, Michalis Vazirgiannis 1,2, El Houcine Bergou 3

4 citations · 43 references · NeurIPS

α

Published on arXiv

2510.22652

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Selecting appropriate weight initialization improves GNN adversarial robustness by up to 50% compared to alternative initialization strategies across diverse models and adversarial attacks.


Graph Neural Networks (GNNs) have demonstrated remarkable performance across a spectrum of graph-related tasks, however concerns persist regarding their vulnerability to adversarial perturbations. While prevailing defense strategies focus primarily on pre-processing techniques and adaptive message-passing schemes, this study delves into an under-explored dimension: the impact of weight initialization and associated hyper-parameters, such as training epochs, on a model's robustness. We introduce a theoretical framework bridging the connection between initialization strategies and a network's resilience to adversarial perturbations. Our analysis reveals a direct relationship between initial weights, number of training epochs and the model's vulnerability, offering new insights into adversarial robustness beyond conventional defense mechanisms. While our primary focus is on GNNs, we extend our theoretical framework, providing a general upper-bound applicable to Deep Neural Networks. Extensive experiments, spanning diverse models and real-world datasets subjected to various adversarial attacks, validate our findings. We illustrate that selecting appropriate initialization not only ensures performance on clean datasets but also enhances model robustness against adversarial perturbations, with observed gaps of up to 50\% compared to alternative initialization approaches.


Key Contributions

  • Theoretical upper bound connecting weight initialization norms and training epochs to adversarial robustness in GCNs and GINs
  • Demonstration that appropriate weight initialization can improve GNN robustness against adversarial attacks by up to 50% without degrading clean accuracy
  • Extension of the theoretical framework to general Deep Neural Networks beyond GNNs

🛡️ Threat Analysis

Input Manipulation Attack

The paper directly addresses adversarial perturbations on GNNs — structural attacks (edge additions/deletions) and node feature attacks that cause misclassification at inference time. The proposed defense (principled weight initialization) is evaluated against benchmark adversarial attacks (PGD, Nettack, Metattack), making this a defense against input manipulation attacks.


Details

Domains
graph
Model Types
gnn
Threat Tags
white_boxblack_boxinference_timedigital
Datasets
CoraCiteseer
Applications
node classificationgraph classification