attack 2025

A Practical Adversarial Attack against Sequence-based Deep Learning Malware Classifiers

Kai Tan , Dongyang Zhan , Lin Ye , Hongli Zhang , Binxing Fang

0 citations

α

Published on arXiv

2509.11836

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Achieves high attack success rate against sequence-based anomaly detection models while preserving malware functionality and satisfying real-world behavioral constraints.

DQN-based adversarial malware perturbation with heuristic backtracking

Novel technique introduced


Sequence-based deep learning models (e.g., RNNs), can detect malware by analyzing its behavioral sequences. Meanwhile, these models are susceptible to adversarial attacks. Attackers can create adversarial samples that alter the sequence characteristics of behavior sequences to deceive malware classifiers. The existing methods for generating adversarial samples typically involve deleting or replacing crucial behaviors in the original data sequences, or inserting benign behaviors that may violate the behavior constraints. However, these methods that directly manipulate sequences make adversarial samples difficult to implement or apply in practice. In this paper, we propose an adversarial attack approach based on Deep Q-Network and a heuristic backtracking search strategy, which can generate perturbation sequences that satisfy practical conditions for successful attacks. Subsequently, we utilize a novel transformation approach that maps modifications back to the source code, thereby avoiding the need to directly modify the behavior log sequences. We conduct an evaluation of our approach, and the results confirm its effectiveness in generating adversarial samples from real-world malware behavior sequences, which have a high success rate in evading anomaly detection models. Furthermore, our approach is practical and can generate adversarial samples while maintaining the functionality of the modified software.


Key Contributions

  • DQN-based adversarial perturbation generator that selects optimal benign behavior insertions to minimally perturb malware behavioral sequences and evade RNN-based classifiers
  • Heuristic backtracking search strategy ensuring generated perturbation sequences satisfy practical behavioral constraints (e.g., valid API call ordering)
  • Novel transformation approach that maps sequence-level modifications back to malware source code, enabling practical real-world attacks without directly tampering with behavior logs

🛡️ Threat Analysis

Input Manipulation Attack

Proposes a method to craft adversarial inputs (modified malware behavioral sequences) that cause misclassification by sequence-based deep learning malware detectors at inference time — core adversarial example attack, using RL (DQN) rather than gradients to optimize perturbations.


Details

Domains
timeseries
Model Types
rnnrl
Threat Tags
black_boxinference_timetargeteddigital
Applications
malware detectionbehavioral sequence classificationanomaly detection