defense 2025

Towards Privacy-Preserving and Heterogeneity-aware Split Federated Learning via Probabilistic Masking

Xingchen Wang 1, Feijie Wu 1, Chenglin Miao 2, Tianchun Li 1, Haoyu Hu 1, Qiming Cao 1, Jing Gao 1, Lu Su 1

0 citations

α

Published on arXiv

2509.14603

Model Inversion Attack

OWASP ML Top 10 — ML03

Key Finding

PM-SFL consistently improves accuracy, communication efficiency, and robustness to data reconstruction attacks compared to noise-injection baselines, especially under data and system heterogeneity

PM-SFL

Novel technique introduced


Split Federated Learning (SFL) has emerged as an efficient alternative to traditional Federated Learning (FL) by reducing client-side computation through model partitioning. However, exchanging of intermediate activations and model updates introduces significant privacy risks, especially from data reconstruction attacks that recover original inputs from intermediate representations. Existing defenses using noise injection often degrade model performance. To overcome these challenges, we present PM-SFL, a scalable and privacy-preserving SFL framework that incorporates Probabilistic Mask training to add structured randomness without relying on explicit noise. This mitigates data reconstruction risks while maintaining model utility. To address data heterogeneity, PM-SFL employs personalized mask learning that tailors submodel structures to each client's local data. For system heterogeneity, we introduce a layer-wise knowledge compensation mechanism, enabling clients with varying resources to participate effectively under adaptive model splitting. Theoretical analysis confirms its privacy protection, and experiments on image and wireless sensing tasks demonstrate that PM-SFL consistently improves accuracy, communication efficiency, and robustness to privacy attacks, with particularly strong performance under data and system heterogeneity.


Key Contributions

  • Probabilistic Mask (PM) training framework that adds structured randomness to intermediate activations to resist data reconstruction attacks without explicit noise injection
  • Personalized mask learning per client to handle data heterogeneity while maintaining model utility
  • Layer-wise knowledge compensation mechanism enabling participation of resource-constrained clients under adaptive model splitting

🛡️ Threat Analysis

Model Inversion Attack

The paper explicitly defends against 'data reconstruction attacks that recover original inputs from intermediate representations' in Split Federated Learning — an adversary observing shared activations tries to reconstruct training data. PM-SFL's probabilistic masking is evaluated against these reconstruction attacks, fitting ML03 precisely.


Details

Domains
visionfederated-learning
Model Types
federatedcnn
Threat Tags
training_timewhite_box
Datasets
CIFAR-10CIFAR-100
Applications
split federated learningimage classificationwireless sensing