Automating Agent Hijacking via Structural Template Injection
Xinhao Deng 1,2, Jiaqing Wu 1,2, Miao Chen 1,3, Yue Xiao 1, Ke Xu 1, Qi Li 1
Published on arXiv
2602.16958
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Phantom significantly outperforms existing baseline prompt injection attacks in attack success rate and query efficiency across Qwen, GPT, and Gemini, with 70+ confirmed real-world vulnerabilities in commercial products.
Phantom / Structural Template Injection (STI)
Novel technique introduced
Agent hijacking, highlighted by OWASP as a critical threat to the Large Language Model (LLM) ecosystem, enables adversaries to manipulate execution by injecting malicious instructions into retrieved content. Most existing attacks rely on manually crafted, semantics-driven prompt manipulation, which often yields low attack success rates and limited transferability to closed-source commercial models. In this paper, we propose Phantom, an automated agent hijacking framework built upon Structured Template Injection that targets the fundamental architectural mechanisms of LLM agents. Our key insight is that agents rely on specific chat template tokens to separate system, user, assistant, and tool instructions. By injecting optimized structured templates into the retrieved context, we induce role confusion and cause the agent to misinterpret the injected content as legitimate user instructions or prior tool outputs. To enhance attack transferability against black-box agents, Phantom introduces a novel attack template search framework. We first perform multi-level template augmentation to increase structural diversity and then train a Template Autoencoder (TAE) to embed discrete templates into a continuous, searchable latent space. Subsequently, we apply Bayesian optimization to efficiently identify optimal adversarial vectors that are decoded into high-potency structured templates. Extensive experiments on Qwen, GPT, and Gemini demonstrate that our framework significantly outperforms existing baselines in both Attack Success Rate (ASR) and query efficiency. Moreover, we identified over 70 vulnerabilities in real-world commercial products that have been confirmed by vendors, underscoring the practical severity of structured template-based hijacking and providing an empirical foundation for securing next-generation agentic systems.
Key Contributions
- Phantom: automated agent hijacking framework using Structural Template Injection (STI) that exploits chat template tokens (system/user/assistant/tool delimiters) to induce role confusion in LLM agents processing retrieved content
- Template Autoencoder (TAE) + Bayesian optimization pipeline that embeds discrete structural templates into a continuous latent space for efficient black-box transferable adversarial template search
- Empirical discovery and vendor-confirmed disclosure of 70+ real-world vulnerabilities in commercial LLM-based products, demonstrating practical severity of STI attacks