InkDrop: Invisible Backdoor Attacks Against Dataset Condensation
He Yang , Dongyi Lv , Song Ma , Wei Xi 1, Zhi Wang 2, Hanlin Gu , Yajie Wang
Published on arXiv
2603.28092
Model Poisoning
OWASP ML Top 10 — ML10
Key Finding
Embeds effective backdoors in condensed datasets with enhanced stealthiness compared to existing attacks while preserving model utility and attack success rates
InkDrop
Novel technique introduced
Dataset Condensation (DC) is a data-efficient learning paradigm that synthesizes small yet informative datasets, enabling models to match the performance of full-data training. However, recent work exposes a critical vulnerability of DC to backdoor attacks, where malicious patterns (\textit{e.g.}, triggers) are implanted into the condensation dataset, inducing targeted misclassification on specific inputs. Existing attacks always prioritize attack effectiveness and model utility, overlooking the crucial dimension of stealthiness. To bridge this gap, we propose InkDrop, which enhances the imperceptibility of malicious manipulation without degrading attack effectiveness and model utility. InkDrop leverages the inherent uncertainty near model decision boundaries, where minor input perturbations can induce semantic shifts, to construct a stealthy and effective backdoor attack. Specifically, InkDrop first selects candidate samples near the target decision boundary that exhibit latent semantic affinity to the target class. It then learns instance-dependent perturbations constrained by perceptual and spatial consistency, embedding targeted malicious behavior into the condensed dataset. Extensive experiments across diverse datasets validate the overall effectiveness of InkDrop, demonstrating its ability to integrate adversarial intent into condensed datasets while preserving model utility and minimizing detectability. Our code is available at https://github.com/lvdongyi/InkDrop.
Key Contributions
- Novel stealthy backdoor attack (InkDrop) that exploits decision boundary uncertainty to embed imperceptible triggers in condensed datasets
- Instance-dependent perturbation learning with perceptual and spatial consistency constraints to minimize detectability
- Demonstrates effective backdoor injection while preserving model utility and visual imperceptibility on condensed datasets
🛡️ Threat Analysis
Paper proposes a backdoor attack (InkDrop) that embeds hidden malicious triggers into condensed datasets, causing targeted misclassification when triggers are present at inference time while maintaining normal behavior otherwise.