survey 2025

On the Security and Privacy of Federated Learning: A Survey with Attacks, Defenses, Frameworks, Applications, and Future Directions

Daniel M. Jimenez-Gutierrez , Yelizaveta Falkouskaya , Jose L. Hernandez-Ramos , Aris Anagnostopoulos , Ioannis Chatzigiannakis , Andrea Vitaletti

0 citations

α

Published on arXiv

2508.13730

Data Poisoning Attack

OWASP ML Top 10 — ML02

Model Poisoning

OWASP ML Top 10 — ML10

Model Inversion Attack

OWASP ML Top 10 — ML03

Key Finding

Identifies non-IID data distribution and the privacy-security-utility trade-off as the most critical open challenges limiting practical deployment of secure FL systems.


Federated Learning (FL) is an emerging distributed machine learning paradigm enabling multiple clients to train a global model collaboratively without sharing their raw data. While FL enhances data privacy by design, it remains vulnerable to various security and privacy threats. This survey provides a comprehensive overview of more than 200 papers regarding the state-of-the-art attacks and defense mechanisms developed to address these challenges, categorizing them into security-enhancing and privacy-preserving techniques. Security-enhancing methods aim to improve FL robustness against malicious behaviors such as byzantine attacks, poisoning, and Sybil attacks. At the same time, privacy-preserving techniques focus on protecting sensitive data through cryptographic approaches, differential privacy, and secure aggregation. We critically analyze the strengths and limitations of existing methods, highlight the trade-offs between privacy, security, and model performance, and discuss the implications of non-IID data distributions on the effectiveness of these defenses. Furthermore, we identify open research challenges and future directions, including the need for scalable, adaptive, and energy-efficient solutions operating in dynamic and heterogeneous FL environments. Our survey aims to guide researchers and practitioners in developing robust and privacy-preserving FL systems, fostering advancements safeguarding collaborative learning frameworks' integrity and confidentiality.


Key Contributions

  • Comprehensive taxonomy of 200+ FL attack and defense papers, categorized into security-enhancing (Byzantine robustness, poisoning, Sybil) and privacy-preserving (DP, secure aggregation, cryptographic) techniques
  • Critical analysis of trade-offs between privacy, security, and model utility under non-IID data distributions
  • Survey of FL frameworks, real-world applications, and open research challenges including scalability and energy efficiency in heterogeneous environments

🛡️ Threat Analysis

Data Poisoning Attack

Extensively covers Byzantine attacks and data/model poisoning in FL — malicious clients corrupting the global model — along with robust aggregation defenses (a primary focus of the security section).

Model Inversion Attack

The privacy section covers gradient leakage/inversion attacks and defenses including secure aggregation and differential privacy specifically to prevent reconstruction of participants' training data from shared gradients.

Model Poisoning

FL backdoor/trojan attacks are a standard major topic in FL security surveys; poisoning with trigger-based hidden behavior is distinct from general Byzantine degradation and warrants its own tag.


Details

Domains
federated-learning
Model Types
federated
Threat Tags
training_timewhite_boxblack_box
Applications
federated learning systemsdistributed machine learning