defense 2026

SpooFL: Spoofing Federated Learning

Isaac Baglin , Xiatian Zhu , Simon Hadfield

0 citations · 29 references · arXiv

α

Published on arXiv

2601.15055

Model Inversion Attack

OWASP ML Top 10 — ML03

Key Finding

SpooFL successfully misdirects gradient inversion attackers into recovering convincing but semantically irrelevant synthetic samples while preserving federated learning model performance

SpooFL

Novel technique introduced


Traditional defenses against Deep Leakage (DL) attacks in Federated Learning (FL) primarily focus on obfuscation, introducing noise, transformations or encryption to degrade an attacker's ability to reconstruct private data. While effective to some extent, these methods often still leak high-level information such as class distributions or feature representations, and are frequently broken by increasingly powerful denoising attacks. We propose a fundamentally different perspective on FL defense: framing it as a spoofing problem.We introduce SpooFL (Figure 1), a spoofing-based defense that deceives attackers into believing they have recovered the true training data, while actually providing convincing but entirely synthetic samples from an unrelated task. Unlike prior synthetic-data defenses that share classes or distributions with the private data and thus still leak semantic information, SpooFL uses a state-of-the-art generative model trained on an external dataset with no class overlap. As a result, attackers are misled into recovering plausible yet completely irrelevant samples, preventing meaningful data leakage while preserving FL training integrity. We implement the first example of such a spoofing defense, and evaluate our method against state-of-the-art DL defenses and demonstrate that it successfully misdirects attackers without compromising model performance significantly.


Key Contributions

  • Reframes FL privacy defense as a spoofing problem rather than obfuscation, causing attackers to converge on plausible but entirely irrelevant synthetic data
  • Uses a generative model trained on an external dataset with zero class overlap to produce convincing decoy gradients, preventing semantic leakage of private data
  • First implementation and evaluation of a spoofing-based defense against state-of-the-art Deep Leakage attacks in federated learning

🛡️ Threat Analysis

Model Inversion Attack

SpooFL defends against Deep Leakage attacks — gradient inversion attacks where an adversary reconstructs participants' private training data from shared gradients in federated learning. The paper's primary contribution is a defense evaluated against data reconstruction attacks.


Details

Domains
federated-learningvision
Model Types
federatedgenerativecnn
Threat Tags
white_boxtraining_time
Applications
federated learningprivacy-preserving machine learningimage classification