attack 2025

MAUI: Reconstructing Private Client Data in Federated Transfer Learning

Ahaan Dabholkar , Atul Sharma , Z. Berkay Celik , Saurabh Bagchi

0 citations

α

Published on arXiv

2509.11451

Model Inversion Attack

OWASP ML Top 10 — ML03

Key Finding

Achieves 40–120% higher PSNR than prior data reconstruction attacks while remaining stealthy, using only classification head gradients across multiple architectures and batch sizes.

MAUI

Novel technique introduced


Recent works in federated learning (FL) have shown the utility of leveraging transfer learning for balancing the benefits of FL and centralized learning. In this setting, federated training happens after a stable point has been reached through conventional training. Global model weights are first centrally pretrained by the server on a public dataset following which only the last few linear layers (the classification head) of the model are finetuned across clients. In this scenario, existing data reconstruction attacks (DRAs) in FL show two key weaknesses. First, strongly input-correlated gradient information from the initial model layers is never shared, significantly degrading reconstruction accuracy. Second, DRAs in which the server makes highly specific, handcrafted manipulations to the model structure or parameters (for e.g., layers with all zero weights, identity mappings and rows with identical weight patterns) are easily detectable by an active client. Improving on these, we propose MAUI, a stealthy DRA that does not require any overt manipulations to the model architecture or weights, and relies solely on the gradients of the classification head. MAUI first extracts "robust" feature representations of the input batch from the gradients of the classification head and subsequently inverts these representations to the original inputs. We report highly accurate reconstructions on the CIFAR10 and ImageNet datasets on a variety of model architectures including convolution networks (CNN, VGG11), ResNets (18, 50), ShuffleNet-V2 and Vision Transformer (ViT B-32), regardless of the batch size. MAUI significantly outperforms prior DRAs in reconstruction quality, achieving 40-120% higher PSNR scores.


Key Contributions

  • MAUI, a stealthy data reconstruction attack requiring no detectable model manipulations, operating solely on classification head gradients in federated transfer learning
  • A two-stage pipeline that extracts robust feature representations from classification head gradients and inverts them back to the original input images
  • Demonstrated 40–120% higher PSNR over prior DRAs across diverse architectures (CNN, VGG11, ResNet-18/50, ShuffleNet-V2, ViT B-32) at arbitrary batch sizes on CIFAR-10 and ImageNet

🛡️ Threat Analysis

Model Inversion Attack

MAUI is a gradient leakage/reconstruction attack in federated learning where a malicious server reconstructs clients' private training data solely from the gradients of the shared classification head — the canonical ML03 threat of adversarial data reconstruction from model gradients in FL.


Details

Domains
visionfederated-learning
Model Types
cnntransformerfederated
Threat Tags
white_boxtraining_timetargeted
Datasets
CIFAR-10ImageNet
Applications
federated learningimage classification