attack 2025

FedPoisonTTP: A Threat Model and Poisoning Attack for Federated Test-Time Personalization

Md Akil Raihan Iftee 1, Syed Md. Ahnaf Hasan 1, Amin Ahsan Ali 1, AKM Mahbubur Rahman 1, Sajib Mistry 2, Aneesh Krishna 2

0 citations · 50 references · arXiv

α

Published on arXiv

2511.19248

Data Poisoning Attack

OWASP ML Top 10 — ML02

Key Finding

Compromised federated participants using FedPoisonTTP substantially diminish overall test-time adaptation performance across clients on corrupted vision benchmarks.

FedPoisonTTP

Novel technique introduced


Test-time personalization in federated learning enables models at clients to adjust online to local domain shifts, enhancing robustness and personalization in deployment. Yet, existing federated learning work largely overlooks the security risks that arise when local adaptation occurs at test time. Heterogeneous domain arrivals, diverse adaptation algorithms, and limited cross-client visibility create vulnerabilities where compromised participants can craft poisoned inputs and submit adversarial updates that undermine both global and per-client performance. To address this threat, we introduce FedPoisonTTP, a realistic grey-box attack framework that explores test-time data poisoning in the federated adaptation setting. FedPoisonTTP distills a surrogate model from adversarial queries, synthesizes in-distribution poisons using feature-consistency, and optimizes attack objectives to generate high-entropy or class-confident poisons that evade common adaptation filters. These poisons are injected during local adaptation and spread through collaborative updates, leading to broad degradation. Extensive experiments on corrupted vision benchmarks show that compromised participants can substantially diminish overall test-time performance.


Key Contributions

  • Identifies and formalizes the security threat of test-time data poisoning in the federated personalization setting, a previously overlooked vulnerability.
  • Proposes FedPoisonTTP, a grey-box attack that distills a surrogate model via adversarial queries and synthesizes in-distribution poisons using feature-consistency constraints.
  • Demonstrates that high-entropy and class-confident poisons evade common adaptation filters and cause broad performance degradation across clients via collaborative update propagation.

🛡️ Threat Analysis

Data Poisoning Attack

FedPoisonTTP is fundamentally a data poisoning attack: it synthesizes poisoned inputs injected during local test-time adaptation (which acts as on-the-fly training data for the adaptation step), corrupting local models that then spread the degradation through federated aggregation. The goal is broad performance degradation — characteristic of ML02 — not a hidden triggered behavior (ML10). The compromised participant submitting adversarial model updates is the canonical federated poisoning pattern that ML02 covers.


Details

Domains
federated-learningvision
Model Types
federated
Threat Tags
grey_boxinference_timeuntargeted
Datasets
corrupted vision benchmarks (e.g., CIFAR-10-C, ImageNet-C)
Applications
federated learningtest-time adaptationdomain generalization