benchmark 2025

Learning-Based Testing for Deep Learning: Enhancing Model Robustness with Adversarial Input Prioritization

Sheikh Md Mushfiqur Rahman , Nasir Eisty

1 citations · 55 references · arXiv

α

Published on arXiv

2509.23961

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

The proposed LBT method consistently outperforms coverage-based and confidence-based baselines in prioritizing fault-revealing adversarial inputs and uncovering all potential faults faster across diverse datasets, architectures, and adversarial attack techniques.

Learning-Based Testing (LBT) for adversarial input prioritization

Novel technique introduced


Context: Deep Neural Networks (DNNs) are increasingly deployed in critical applications, where resilience against adversarial inputs is paramount. However, whether coverage-based or confidence-based, existing test prioritization methods often fail to efficiently identify the most fault-revealing inputs, limiting their practical effectiveness. Aims: This project aims to enhance fault detection and model robustness in DNNs by integrating Learning-Based Testing (LBT) with hypothesis and mutation testing to efficiently prioritize adversarial test cases. Methods: Our method selects a subset of adversarial inputs with a high likelihood of exposing model faults, without relying on architecture-specific characteristics or formal verification, making it adaptable across diverse DNNs. Results: Our results demonstrate that the proposed LBT method consistently surpasses baseline approaches in prioritizing fault-revealing inputs and accelerating fault detection. By efficiently organizing test permutations, it uncovers all potential faults significantly faster across various datasets, model architectures, and adversarial attack techniques. Conclusion: Beyond improving fault detection, our method preserves input diversity and provides effective guidance for model retraining, further enhancing robustness. These advantages establish our approach as a powerful and practical solution for adversarial test prioritization in real-world DNN applications.


Key Contributions

  • An LBT-based adversarial test prioritization method that selects fault-revealing inputs without requiring access to model internals or architecture-specific knowledge
  • Integration of hypothesis and mutation testing into the adversarial input selection loop to accelerate fault detection across diverse DNN architectures
  • Retraining guidance derived from prioritized adversarial inputs that preserves input diversity while enhancing model robustness

🛡️ Threat Analysis

Input Manipulation Attack

The paper's primary goal is improving DNN robustness against adversarial inputs — the LBT method selects the most fault-revealing adversarial test cases and provides retraining guidance, making it a defense-oriented adversarial testing methodology targeting inference-time input manipulation threats.


Details

Domains
vision
Model Types
cnntransformer
Threat Tags
white_boxblack_boxinference_time
Datasets
CIFAR-10ImageNet
Applications
image classificationautonomous drivingsafety-critical dnn systems