GraphDLG: Exploring Deep Leakage from Gradients in Federated Graph Learning
Shuyue Wei 1, Wantong Chen 2, Tongyu Wei 2, Chen Gong 3, Yongxin Tong 2, Lizhen Cui 1
Published on arXiv
2601.19745
Model Inversion Attack
OWASP ML Top 10 — ML03
Key Finding
GraphDLG outperforms prior DLG methods by over 5.46% in node feature reconstruction (MSE) and over 25.04% in graph structure reconstruction (AUC) across federated GNN settings.
GraphDLG
Novel technique introduced
Federated graph learning (FGL) has recently emerged as a promising privacy-preserving paradigm that enables distributed graph learning across multiple data owners. A critical privacy concern in federated learning is whether an adversary can recover raw data from shared gradients, a vulnerability known as deep leakage from gradients (DLG). However, most prior studies on the DLG problem focused on image or text data, and it remains an open question whether graphs can be effectively recovered, particularly when the graph structure and node features are uniquely entangled in GNNs. In this work, we first theoretically analyze the components in FGL and derive a crucial insight: once the graph structure is recovered, node features can be obtained through a closed-form recursive rule. Building on this analysis, we propose GraphDLG, a novel approach to recover raw training graphs from shared gradients in FGL, which can utilize randomly generated graphs or client-side training graphs as auxiliaries to enhance recovery. Extensive experiments demonstrate that GraphDLG outperforms existing solutions by successfully decoupling the graph structure and node features, achieving improvements of over 5.46% (by MSE) for node feature reconstruction and over 25.04% (by AUC) for graph structure reconstruction.
Key Contributions
- Theoretical analysis showing that once graph structure is recovered from gradients, node features can be obtained through a closed-form recursive rule, decoupling the joint recovery problem.
- GraphDLG attack framework that reconstructs raw training graphs from shared gradients in federated GNNs, optionally leveraging randomly generated or client-side auxiliary graphs to improve recovery.
- Empirical demonstration that GraphDLG outperforms existing DLG methods by >5.46% (MSE) on node feature reconstruction and >25.04% (AUC) on graph structure reconstruction.
🛡️ Threat Analysis
GraphDLG is a gradient leakage/reconstruction attack in federated learning where an adversary reconstructs raw training graphs (structure + node features) from gradients shared by clients during federated GNN training — canonical ML03 gradient inversion attack extended to graph-structured data.