defense 2025

Pigeon-SL: Robust Split Learning Framework for Edge Intelligence under Malicious Clients

Sangjun Park 1, Tony Q.S. Quek 2, Hyowoon Seo 3

0 citations

α

Published on arXiv

2508.02235

Data Poisoning Attack

OWASP ML Top 10 — ML02

Key Finding

Pigeon-SL guarantees isolation of malicious updates via the pigeonhole principle and demonstrates significant accuracy improvements over baseline split learning under label flipping, activation manipulation, and gradient manipulation attacks.

Pigeon-SL

Novel technique introduced


Recent advances in split learning (SL) have established it as a promising framework for privacy-preserving, communication-efficient distributed learning at the network edge. However, SL's sequential update process is vulnerable to even a single malicious client, which can significantly degrade model accuracy. To address this, we introduce Pigeon-SL, a novel scheme grounded in the pigeonhole principle that guarantees at least one entirely honest cluster among M clients, even when up to N of them are adversarial. In each global round, the access point partitions the clients into N+1 clusters, trains each cluster independently via vanilla SL, and evaluates their validation losses on a shared dataset. Only the cluster with the lowest loss advances, thereby isolating and discarding malicious updates. We further enhance training and communication efficiency with Pigeon-SL+, which repeats training on the selected cluster to match the update throughput of standard SL. We validate the robustness and effectiveness of our approach under three representative attack models -- label flipping, activation and gradient manipulation -- demonstrating significant improvements in accuracy and resilience over baseline SL methods in future intelligent wireless networks.


Key Contributions

  • Pigeon-SL: a pigeonhole-principle-based client clustering scheme that guarantees at least one fully honest cluster among M clients when up to N are adversarial, selecting only the lowest-validation-loss cluster each round
  • Pigeon-SL+: an efficiency enhancement that repeats training on the selected cluster to match the update throughput of standard split learning
  • Empirical validation against three attack models (label flipping, activation manipulation, gradient manipulation) demonstrating improved accuracy and resilience over baseline SL

🛡️ Threat Analysis

Data Poisoning Attack

Pigeon-SL defends against malicious clients in split learning who perform label flipping (data poisoning) and activation/gradient manipulation (Byzantine-style model update corruption) to degrade global model accuracy — all classic ML02 attack vectors. The defense (cluster selection via validation loss) is a form of robust aggregation against poisoning.


Details

Domains
federated-learning
Model Types
federated
Threat Tags
training_timegrey_boxuntargeted
Applications
split learningedge intelligencedistributed learning