attack 2026

InkDrop: Invisible Backdoor Attacks Against Dataset Condensation

He Yang , Dongyi Lv , Song Ma , Wei Xi 1, Zhi Wang 2, Hanlin Gu , Yajie Wang

0 citations

α

Published on arXiv

2603.28092

Model Poisoning

OWASP ML Top 10 — ML10

Key Finding

Embeds effective backdoors in condensed datasets with enhanced stealthiness compared to existing attacks while preserving model utility and attack success rates

InkDrop

Novel technique introduced


Dataset Condensation (DC) is a data-efficient learning paradigm that synthesizes small yet informative datasets, enabling models to match the performance of full-data training. However, recent work exposes a critical vulnerability of DC to backdoor attacks, where malicious patterns (\textit{e.g.}, triggers) are implanted into the condensation dataset, inducing targeted misclassification on specific inputs. Existing attacks always prioritize attack effectiveness and model utility, overlooking the crucial dimension of stealthiness. To bridge this gap, we propose InkDrop, which enhances the imperceptibility of malicious manipulation without degrading attack effectiveness and model utility. InkDrop leverages the inherent uncertainty near model decision boundaries, where minor input perturbations can induce semantic shifts, to construct a stealthy and effective backdoor attack. Specifically, InkDrop first selects candidate samples near the target decision boundary that exhibit latent semantic affinity to the target class. It then learns instance-dependent perturbations constrained by perceptual and spatial consistency, embedding targeted malicious behavior into the condensed dataset. Extensive experiments across diverse datasets validate the overall effectiveness of InkDrop, demonstrating its ability to integrate adversarial intent into condensed datasets while preserving model utility and minimizing detectability. Our code is available at https://github.com/lvdongyi/InkDrop.


Key Contributions

  • Novel stealthy backdoor attack (InkDrop) that exploits decision boundary uncertainty to embed imperceptible triggers in condensed datasets
  • Instance-dependent perturbation learning with perceptual and spatial consistency constraints to minimize detectability
  • Demonstrates effective backdoor injection while preserving model utility and visual imperceptibility on condensed datasets

🛡️ Threat Analysis

Model Poisoning

Paper proposes a backdoor attack (InkDrop) that embeds hidden malicious triggers into condensed datasets, causing targeted misclassification when triggers are present at inference time while maintaining normal behavior otherwise.


Details

Domains
vision
Model Types
cnn
Threat Tags
training_timetargeteddigital
Applications
dataset condensationimage classification