The Shawshank Redemption of Embodied AI: Understanding and Benchmarking Indirect Environmental Jailbreaks
Chunyang Li 1, Zifeng Kang 2, Junwei Zhang 1, Zhuo Ma 1, Anda Cheng 3, Xinghua Li 1, Jianfeng Ma 1
Published on arXiv
2511.16347
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
SHAWSHANK outperforms all 11 existing jailbreak methods and successfully compromises all six tested VLMs via environment-injected instructions, with current defenses only partially mitigating the attack.
SHAWSHANK / Indirect Environmental Jailbreak (IEJ)
Novel technique introduced
The adoption of Vision-Language Models (VLMs) in embodied AI agents, while being effective, brings safety concerns such as jailbreaking. Prior work have explored the possibility of directly jailbreaking the embodied agents through elaborated multi-modal prompts. However, no prior work has studied or even reported indirect jailbreaks in embodied AI, where a black-box attacker induces a jailbreak without issuing direct prompts to the embodied agent. In this paper, we propose, for the first time, indirect environmental jailbreak (IEJ), a novel attack to jailbreak embodied AI via indirect prompt injected into the environment, such as malicious instructions written on a wall. Our key insight is that embodied AI does not ''think twice'' about the instructions provided by the environment -- a blind trust that attackers can exploit to jailbreak the embodied agent. We further design and implement open-source prototypes of two fully-automated frameworks: SHAWSHANK, the first automatic attack generation framework for the proposed attack IEJ; and SHAWSHANK-FORGE, the first automatic benchmark generation framework for IEJ. Then, using SHAWSHANK-FORGE, we automatically construct SHAWSHANK-BENCH, the first benchmark for indirectly jailbreaking embodied agents. Together, our two frameworks and one benchmark answer the questions of what content can be used for malicious IEJ instructions, where they should be placed, and how IEJ can be systematically evaluated. Evaluation results show that SHAWSHANK outperforms eleven existing methods across 3,957 task-scene combinations and compromises all six tested VLMs. Furthermore, current defenses only partially mitigate our attack, and we have responsibly disclosed our findings to all affected VLM vendors.
Key Contributions
- Proposes Indirect Environmental Jailbreak (IEJ), the first attack that jailbreaks embodied VLM agents by injecting malicious instructions into the physical/virtual environment rather than issuing direct prompts.
- Designs SHAWSHANK, an automated attack generation framework for IEJ, which outperforms 11 baselines across 3,957 task-scene combinations and successfully compromises all 6 tested VLMs.
- Builds SHAWSHANK-FORGE and SHAWSHANK-BENCH, the first automatic benchmark generation framework and benchmark for evaluating indirect embodied AI jailbreaks.