defense 2026

Byzantine-Robust and Differentially Private Federated Optimization under Weaker Assumptions

Rustem Islamov 1, Grigory Malinovsky 2, Alexander Gaponov 2, Aurelien Lucchi 1, Peter Richtárik 2, Eduard Gorbunov 3

0 citations

α

Published on arXiv

2603.23472

Data Poisoning Attack

OWASP ML Top 10 — ML02

Key Finding

Achieves state-of-the-art convergence rates under Byzantine attacks and DP constraints without requiring bounded gradients or auxiliary server datasets

Byz-Clip21-SGD2M

Novel technique introduced


Federated Learning (FL) enables heterogeneous clients to collaboratively train a shared model without centralizing their raw data, offering an inherent level of privacy. However, gradients and model updates can still leak sensitive information, while malicious servers may mount adversarial attacks such as Byzantine manipulation. These vulnerabilities highlight the need to address differential privacy (DP) and Byzantine robustness within a unified framework. Existing approaches, however, often rely on unrealistic assumptions such as bounded gradients, require auxiliary server-side datasets, or fail to provide convergence guarantees. We address these limitations by proposing Byz-Clip21-SGD2M, a new algorithm that integrates robust aggregation with double momentum and carefully designed clipping. We prove high-probability convergence guarantees under standard $L$-smoothness and $σ$-sub-Gaussian gradient noise assumptions, thereby relaxing conditions that dominate prior work. Our analysis recovers state-of-the-art convergence rates in the absence of adversaries and improves utility guarantees under Byzantine and DP settings. Empirical evaluations on CNN and MLP models trained on MNIST further validate the effectiveness of our approach.


Key Contributions

  • Byz-Clip21-SGD2M algorithm combining robust aggregation with double momentum and clipping for Byzantine+DP settings
  • Convergence guarantees under L-smoothness and sub-Gaussian noise without bounded gradient assumptions
  • Unified framework handling both Byzantine robustness and differential privacy with improved utility

🛡️ Threat Analysis

Data Poisoning Attack

Defends against Byzantine attacks in federated learning where malicious clients send arbitrary model updates to degrade global model performance — this is data poisoning via malicious gradient updates.


Details

Domains
federated-learning
Model Types
federatedcnntraditional_ml
Threat Tags
training_time
Datasets
MNIST
Applications
federated learning