benchmark 2025

Phantom Menace: Exploring and Enhancing the Robustness of VLA Models Against Physical Sensor Attacks

Xuancun Lu 1, Jiaxiang Chen 1, Shilin Xiao 1, Zizhi Jin 1, Zhangrui Chen 1, Hanwen Yu 1, Bohan Qian 1, Ruochen Zhou 2, Xiaoyu Ji 1, Wenyuan Xu 1

0 citations · 39 references · arXiv

α

Published on arXiv

2511.10008

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Physical sensor attacks (laser, EMI, ultrasound) significantly degrade VLA model performance with vulnerability patterns tied to task type and architecture, and adversarial training restores robustness against out-of-distribution physical perturbations

Real-Sim-Real

Novel technique introduced


Vision-Language-Action (VLA) models revolutionize robotic systems by enabling end-to-end perception-to-action pipelines that integrate multiple sensory modalities, such as visual signals processed by cameras and auditory signals captured by microphones. This multi-modality integration allows VLA models to interpret complex, real-world environments using diverse sensor data streams. Given the fact that VLA-based systems heavily rely on the sensory input, the security of VLA models against physical-world sensor attacks remains critically underexplored. To address this gap, we present the first systematic study of physical sensor attacks against VLAs, quantifying the influence of sensor attacks and investigating the defenses for VLA models. We introduce a novel "Real-Sim-Real" framework that automatically simulates physics-based sensor attack vectors, including six attacks targeting cameras and two targeting microphones, and validates them on real robotic systems. Through large-scale evaluations across various VLA architectures and tasks under varying attack parameters, we demonstrate significant vulnerabilities, with susceptibility patterns that reveal critical dependencies on task types and model designs. We further develop an adversarial-training-based defense that enhances VLA robustness against out-of-distribution physical perturbations caused by sensor attacks while preserving model performance. Our findings expose an urgent need for standardized robustness benchmarks and mitigation strategies to secure VLA deployments in safety-critical environments.


Key Contributions

  • Real-Sim-Real framework that automatically simulates eight physics-based sensor attacks (six camera, two microphone) and validates findings on real robotic systems
  • Large-scale evaluation of four VLA model architectures across four datasets revealing significant vulnerability patterns dependent on task type and model design
  • Adversarial-training-based defense that improves VLA robustness against out-of-distribution physical sensor perturbations while preserving clean task performance

🛡️ Threat Analysis

Input Manipulation Attack

Physical sensor attacks (laser dazzling, electromagnetic interference, ultrasonic injection) corrupt camera and microphone inputs to VLA models at inference time, causing incorrect action outputs — this is physical adversarial input manipulation. The adversarial-training-based defense is also a canonical ML01 countermeasure.


Details

Domains
multimodalvisionaudio
Model Types
vlmmultimodal
Threat Tags
physicalinference_timeblack_boxuntargeted
Applications
robotic manipulationembodied aivision-language-action systems