FedAgain: A Trust-Based and Robust Federated Learning Strategy for an Automated Kidney Stone Identification in Ureteroscopy
Ivan Reyes-Amezcua 1, Francisco Lopez-Tiro 2,3,4, Clément Larose 3,4, Christian Daul 3, Andres Mendez-Vazquez 1, Gilberto Ochoa-Ruiz 2
Published on arXiv
2603.19512
Data Poisoning Attack
OWASP ML Top 10 — ML02
Key Finding
Consistently outperforms standard federated learning baselines under non-IID data and corrupted-client scenarios while maintaining diagnostic accuracy and stability
FedAgain
Novel technique introduced
The reliability of artificial intelligence (AI) in medical imaging critically depends on its robustness to heterogeneous and corrupted images acquired with diverse devices across different hospitals which is highly challenging. Therefore, this paper introduces FedAgain, a trust-based Federated Learning (Federated Learning) strategy designed to enhance robustness and generalization for automated kidney stone identification from endoscopic images. FedAgain integrates a dual trust mechanism that combines benchmark reliability and model divergence to dynamically weight client contributions, mitigating the impact of noisy or adversarial updates during aggregation. The framework enables the training of collaborative models across multiple institutions while preserving data privacy and promoting stable convergence under real-world conditions. Extensive experiments across five datasets, including two canonical benchmarks (MNIST and CIFAR-10), two private multi-institutional kidney stone datasets, and one public dataset (MyStone), demonstrate that FedAgain consistently outperforms standard Federated Learning baselines under non-identically and independently distributed (non-IID) data and corrupted-client scenarios. By maintaining diagnostic accuracy and performance stability under varying conditions, FedAgain represents a practical advance toward reliable, privacy-preserving, and clinically deployable federated AI for medical imaging.
Key Contributions
- Dual trust mechanism combining benchmark reliability and model divergence to weight client contributions in federated learning
- Robust aggregation strategy that mitigates impact of noisy or adversarial updates while preserving privacy
- Demonstrated robustness under non-IID data and corrupted-client scenarios across medical imaging and benchmark datasets
🛡️ Threat Analysis
FedAgain defends against corrupted/adversarial client updates in federated learning by dynamically weighting client contributions based on benchmark reliability and model divergence — this directly addresses Byzantine attacks and noisy data poisoning during FL aggregation.