ROAST: Risk-aware Outlier-exposure for Adversarial Selective Training of Anomaly Detectors Against Evasion Attacks
Mohammed Elnawawy 1, Gargi Mitra 1, Shahrear Iqbal 2, Karthik Pattabiraman 1
Published on arXiv
2603.26093
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
Increases anomaly detector recall by 16.2% while reducing training time by 88.3% on average, with minimal precision impact
ROAST
Novel technique introduced
Safety-critical domains like healthcare rely on deep neural networks (DNNs) for prediction, yet DNNs remain vulnerable to evasion attacks. Anomaly detectors (ADs) are widely used to protect DNNs, but conventional ADs are trained indiscriminately on benign data from all patients, overlooking physiological differences that introduce noise, degrade robustness, and reduce recall. In this paper, we propose ROAST, a novel risk-aware outlier exposure selective training framework that improves AD recall without sacrificing precision. ROAST identifies patients who are less vulnerable to attack and focuses training on these cleaner, more reliable data, thereby reducing false negatives and improving recall. To preserve precision, the framework applies outlier exposure by injecting adversarial samples into the training set of the less vulnerable patients, avoiding noisy data from others. Experiments show that ROAST increases recall by 16.2\% while reducing the training time by 88.3\% on average compared to indiscriminate training, with minimal impact on precision.
Key Contributions
- Risk-aware selective training framework that identifies less vulnerable patients for cleaner training data
- Outlier exposure technique that injects adversarial samples into training set of reliable data subsets
- Achieves 16.2% recall improvement with 88.3% reduction in training time compared to indiscriminate training
🛡️ Threat Analysis
Primary focus is defending against adversarial evasion attacks at inference time - the anomaly detector is designed to detect adversarial examples that would cause misclassification.