The Easy Path to Robustness: Coreset Selection using Sample Hardness
Pranav Ramesh 1, Arjun Roy 2, Deepak Ravikumar 2, Kaushik Roy 2, Gopalakrishnan Srinivasan 1
Published on arXiv
2510.11018
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
EasyCore achieves up to 7% improvement in adversarial accuracy under standard training and 5% under TRADES compared to existing coreset selection methods.
EasyCore
Novel technique introduced
Designing adversarially robust models from a data-centric perspective requires understanding which input samples are most crucial for learning resilient features. While coreset selection provides a mechanism for efficient training on data subsets, current algorithms are designed for clean accuracy and fall short in preserving robustness. To address this, we propose a framework linking a sample's adversarial vulnerability to its \textit{hardness}, which we quantify using the average input gradient norm (AIGN) over training. We demonstrate that \textit{easy} samples (with low AIGN) are less vulnerable and occupy regions further from the decision boundary. Leveraging this insight, we present EasyCore, a coreset selection algorithm that retains only the samples with low AIGN for training. We empirically show that models trained on EasyCore-selected data achieve significantly higher adversarial accuracy than those trained with competing coreset methods under both standard and adversarial training. As AIGN is a model-agnostic dataset property, EasyCore is an efficient and widely applicable data-centric method for improving adversarial robustness. We show that EasyCore achieves up to 7\% and 5\% improvement in adversarial accuracy under standard training and TRADES adversarial training, respectively, compared to existing coreset methods.
Key Contributions
- Framework linking adversarial vulnerability to sample hardness quantified via average input gradient norm (AIGN) over training
- EasyCore: a coreset selection algorithm that retains only low-AIGN (easy) samples, improving adversarial robustness in a model-agnostic, data-centric way
- Empirical demonstration of up to 7% and 5% adversarial accuracy gains over competing coreset methods under standard and TRADES adversarial training, respectively
🛡️ Threat Analysis
Primary contribution is a defense against adversarial input manipulation attacks — EasyCore improves adversarial accuracy by selecting training data that produces models more resilient to adversarial examples, evaluated under standard and TRADES adversarial training.