RPP: A Certified Poisoned-Sample Detection Framework for Backdoor Attacks under Dataset Imbalance
Miao Lin 1, Feng Yu 2, Rui Ning 1, Lusi Li 1, Jiawei Chen 1, Qian Lou 3, Mengxin Zheng 3, Chunsheng Xin 4, Hongyi Wu 5
Published on arXiv
2602.00183
Model Poisoning
OWASP ML Top 10 — ML10
Key Finding
RPP achieves significantly higher backdoor detection accuracy than 12 state-of-the-art defenses across 10 attack types, with certified guarantees on false positive rate under class-imbalanced conditions.
RPP (Randomized Probability Perturbation)
Novel technique introduced
Deep neural networks are highly susceptible to backdoor attacks, yet most defense methods to date rely on balanced data, overlooking the pervasive class imbalance in real-world scenarios that can amplify backdoor threats. This paper presents the first in-depth investigation of how the dataset imbalance amplifies backdoor vulnerability, showing that (i) the imbalance induces a majority-class bias that increases susceptibility and (ii) conventional defenses degrade significantly as the imbalance grows. To address this, we propose Randomized Probability Perturbation (RPP), a certified poisoned-sample detection framework that operates in a black-box setting using only model output probabilities. For any inspected sample, RPP determines whether the input has been backdoor-manipulated, while offering provable within-domain detectability guarantees and a probabilistic upper bound on the false positive rate. Extensive experiments on five benchmarks (MNIST, SVHN, CIFAR-10, TinyImageNet and ImageNet10) covering 10 backdoor attacks and 12 baseline defenses show that RPP achieves significantly higher detection accuracy than state-of-the-art defenses, particularly under dataset imbalance. RPP establishes a theoretical and practical foundation for defending against backdoor attacks in real-world environments with imbalanced data.
Key Contributions
- First in-depth analysis showing dataset imbalance amplifies backdoor vulnerability via majority-class bias and degrades conventional defenses
- RPP: a certified poisoned-sample detection framework using only black-box model output probabilities, with provable within-domain detectability guarantees and a probabilistic false positive rate upper bound
- Extensive evaluation across 5 benchmarks, 10 backdoor attacks, and 12 baseline defenses demonstrating superior detection accuracy especially under imbalanced settings
🛡️ Threat Analysis
RPP is a backdoor/trojan defense — it detects whether inference-time inputs have been backdoor-manipulated using trigger patterns, evaluated against 10 distinct backdoor attacks with certified detectability guarantees.