SpooFL: Spoofing Federated Learning
Isaac Baglin , Xiatian Zhu , Simon Hadfield
Published on arXiv
2601.15055
Model Inversion Attack
OWASP ML Top 10 — ML03
Key Finding
SpooFL successfully misdirects gradient inversion attackers into recovering convincing but semantically irrelevant synthetic samples while preserving federated learning model performance
SpooFL
Novel technique introduced
Traditional defenses against Deep Leakage (DL) attacks in Federated Learning (FL) primarily focus on obfuscation, introducing noise, transformations or encryption to degrade an attacker's ability to reconstruct private data. While effective to some extent, these methods often still leak high-level information such as class distributions or feature representations, and are frequently broken by increasingly powerful denoising attacks. We propose a fundamentally different perspective on FL defense: framing it as a spoofing problem.We introduce SpooFL (Figure 1), a spoofing-based defense that deceives attackers into believing they have recovered the true training data, while actually providing convincing but entirely synthetic samples from an unrelated task. Unlike prior synthetic-data defenses that share classes or distributions with the private data and thus still leak semantic information, SpooFL uses a state-of-the-art generative model trained on an external dataset with no class overlap. As a result, attackers are misled into recovering plausible yet completely irrelevant samples, preventing meaningful data leakage while preserving FL training integrity. We implement the first example of such a spoofing defense, and evaluate our method against state-of-the-art DL defenses and demonstrate that it successfully misdirects attackers without compromising model performance significantly.
Key Contributions
- Reframes FL privacy defense as a spoofing problem rather than obfuscation, causing attackers to converge on plausible but entirely irrelevant synthetic data
- Uses a generative model trained on an external dataset with zero class overlap to produce convincing decoy gradients, preventing semantic leakage of private data
- First implementation and evaluation of a spoofing-based defense against state-of-the-art Deep Leakage attacks in federated learning
🛡️ Threat Analysis
SpooFL defends against Deep Leakage attacks — gradient inversion attacks where an adversary reconstructs participants' private training data from shared gradients in federated learning. The paper's primary contribution is a defense evaluated against data reconstruction attacks.