CIBER: A Comprehensive Benchmark for Security Evaluation of Code Interpreter Agents
Lei Ba 1, Qinbin Li 2, Songze Li 1
Published on arXiv
2602.19547
Prompt Injection
OWASP LLM Top 10 — LLM01
Excessive Agency
OWASP LLM Top 10 — LLM08
Key Finding
Across six foundation models, natural language-disguised attacks achieve 14.1% higher attack success rate than code-based attacks, and all agents catastrophically fail against implicit semantic hazards despite robust defenses against explicit threats.
CIBER
Novel technique introduced
LLM-based code interpreter agents are increasingly deployed in critical workflows, yet their robustness against risks introduced by their code execution capabilities remains underexplored. Existing benchmarks are limited to static datasets or simulated environments, failing to capture the security risks arising from dynamic code execution, tool interactions, and multi-turn context. To bridge this gap, we introduce CIBER, an automated benchmark that combines dynamic attack generation, isolated secure sandboxing, and state-aware evaluation to systematically assess the vulnerability of code interpreter agents against four major types of adversarial attacks: Direct/Indirect Prompt Injection, Memory Poisoning, and Prompt-based Backdoor. We evaluate six foundation models across two representative code interpreter agents (OpenInterpreter and OpenCodeInterpreter), incorporating a controlled study of identical models. Our results reveal that Interpreter Architecture and Model Alignment Set the Security Baseline. Structural integration enables aligned specialized models to outperform generic SOTA models. Conversely, high intelligence paradoxically increases susceptibility to complex adversarial prompts due to stronger instruction adherence. Furthermore, we identify a "Natural Language Disguise" Phenomenon, where natural language functions as a significantly more effective input modality than explicit code snippets (+14.1% ASR), thereby bypassing syntax-based defenses. Finally, we expose an alarming Security Polarization, where agents exhibit robust defenses against explicit threats yet fail catastrophically against implicit semantic hazards, highlighting a fundamental blind spot in current pattern-matching protection approaches.
Key Contributions
- CIBER: an automated benchmark with dynamic attack generation, dockerized sandboxing with real system privileges, and state-aware evaluation for code interpreter agent security
- Empirical finding that natural language disguise is significantly more effective than explicit code for adversarial attacks (+14.1% ASR), bypassing syntax-based defenses
- Discovery of Security Polarization — agents robustly resist explicit threats but fail catastrophically against implicit semantic hazards — and identification of trusted channels (tool outputs, conversation history) as structural blind spots