defense 2025

The Easy Path to Robustness: Coreset Selection using Sample Hardness

Pranav Ramesh 1, Arjun Roy 2, Deepak Ravikumar 2, Kaushik Roy 2, Gopalakrishnan Srinivasan 1

0 citations · 31 references · arXiv

α

Published on arXiv

2510.11018

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

EasyCore achieves up to 7% improvement in adversarial accuracy under standard training and 5% under TRADES compared to existing coreset selection methods.

EasyCore

Novel technique introduced


Designing adversarially robust models from a data-centric perspective requires understanding which input samples are most crucial for learning resilient features. While coreset selection provides a mechanism for efficient training on data subsets, current algorithms are designed for clean accuracy and fall short in preserving robustness. To address this, we propose a framework linking a sample's adversarial vulnerability to its \textit{hardness}, which we quantify using the average input gradient norm (AIGN) over training. We demonstrate that \textit{easy} samples (with low AIGN) are less vulnerable and occupy regions further from the decision boundary. Leveraging this insight, we present EasyCore, a coreset selection algorithm that retains only the samples with low AIGN for training. We empirically show that models trained on EasyCore-selected data achieve significantly higher adversarial accuracy than those trained with competing coreset methods under both standard and adversarial training. As AIGN is a model-agnostic dataset property, EasyCore is an efficient and widely applicable data-centric method for improving adversarial robustness. We show that EasyCore achieves up to 7\% and 5\% improvement in adversarial accuracy under standard training and TRADES adversarial training, respectively, compared to existing coreset methods.


Key Contributions

  • Framework linking adversarial vulnerability to sample hardness quantified via average input gradient norm (AIGN) over training
  • EasyCore: a coreset selection algorithm that retains only low-AIGN (easy) samples, improving adversarial robustness in a model-agnostic, data-centric way
  • Empirical demonstration of up to 7% and 5% adversarial accuracy gains over competing coreset methods under standard and TRADES adversarial training, respectively

🛡️ Threat Analysis

Input Manipulation Attack

Primary contribution is a defense against adversarial input manipulation attacks — EasyCore improves adversarial accuracy by selecting training data that produces models more resilient to adversarial examples, evaluated under standard and TRADES adversarial training.


Details

Domains
vision
Model Types
cnn
Threat Tags
white_boxinference_timetraining_timedigital
Datasets
CIFAR-10CIFAR-100
Applications
image classification