Adversarial Patch Attacks on Vision-Based Cargo Occupancy Estimation via Differentiable 3D Simulation
Mohamed Rissal Hedna , Sesugh Samuel Nder
Published on arXiv
2511.19254
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
3D-optimized adversarial patches achieve 84.94% attack success on denial-of-service (empty→full) and 30.32% on concealment (full→empty) attacks against a cargo occupancy CNN.
Differentiable 3D Patch Optimization (Mitsuba 3 + EOT)
Novel technique introduced
Computer vision systems are increasingly adopted in modern logistics operations, including the estimation of trailer occupancy for planning, routing, and billing. Although effective, such systems may be vulnerable to physical adversarial attacks, particularly adversarial patches that can be printed and placed on interior surfaces. In this work, we study the feasibility of such attacks on a convolutional cargo-occupancy classifier using fully simulated 3D environments. Using Mitsuba 3 for differentiable rendering, we optimize patch textures across variations in geometry, lighting, and viewpoint, and compare their effectiveness to a 2D compositing baseline. Our experiments demonstrate that 3D-optimized patches achieve high attack success rates, especially in a denial-of-service scenario (empty to full), where success reaches 84.94 percent. Concealment attacks (full to empty) prove more challenging but still reach 30.32 percent. We analyze the factors influencing attack success, discuss implications for the security of automated logistics pipelines, and highlight directions for strengthening physical robustness. To our knowledge, this is the first study to investigate adversarial patch attacks for cargo-occupancy estimation in physically realistic, fully simulated 3D scenes.
Key Contributions
- Differentiable 3D simulation pipeline using Mitsuba 3 for adversarial patch optimization in cargo trailer environments, with surface-aware patch placement sampling.
- Comparison of 2D compositing vs. 3D scene-space patch optimization across 5,910 rendered scenes, demonstrating 3D optimization superiority.
- Systematic evaluation of two attack scenarios — denial-of-service (empty→full, 84.94%) and concealment (full→empty, 30.32%) — against a logistics occupancy classifier.
🛡️ Threat Analysis
Proposes physical adversarial patch attacks (printable, placeable in real trailer interiors) that cause targeted misclassification of a CNN occupancy classifier at inference time — core adversarial patch / physical adversarial example attack falling directly under ML01.