On Robustness of Linear Classifiers to Targeted Data Poisoning
Nakshatra Gupta 1, Sumanth Prabhu 2, Supratik Chakraborty 3, R Venkatesh 1
Published on arXiv
2511.12722
Data Poisoning Attack
OWASP ML Top 10 — ML02
Key Finding
Robustness bounds are efficiently computable in practice for publicly available datasets, succeeding in many cases where state-of-the-art methods fail, and poisoning exceeding the bounds reliably alters the targeted test point's classification.
Data poisoning is a training-time attack that undermines the trustworthiness of learned models. In a targeted data poisoning attack, an adversary manipulates the training dataset to alter the classification of a targeted test point. Given the typically large size of training dataset, manual detection of poisoning is difficult. An alternative is to automatically measure a dataset's robustness against such an attack, which is the focus of this paper. We consider a threat model wherein an adversary can only perturb the labels of the training dataset, with knowledge limited to the hypothesis space of the victim's model. In this setting, we prove that finding the robustness is an NP-Complete problem, even when hypotheses are linear classifiers. To overcome this, we present a technique that finds lower and upper bounds of robustness. Our implementation of the technique computes these bounds efficiently in practice for many publicly available datasets. We experimentally demonstrate the effectiveness of our approach. Specifically, a poisoning exceeding the identified robustness bounds significantly impacts test point classification. We are also able to compute these bounds in many more cases where state-of-the-art techniques fail.
Key Contributions
- Proves that computing exact robustness against targeted label-flipping poisoning is NP-Complete even for linear classifiers
- Presents a practical technique computing lower and upper bounds on dataset robustness against targeted data poisoning
- Empirically demonstrates that poisoning beyond the identified bounds reliably flips the targeted test point's classification
🛡️ Threat Analysis
Directly addresses targeted data poisoning via label-flipping attacks at training time, providing formal robustness bounds against an adversary who manipulates training labels to alter classification of a specific test point.