Supply-Chain Poisoning Attacks Against LLM Coding Agent Skill Ecosystems
Yubin Qu 1, Yi Liu 2,3,4, Tongcheng Geng 1, Gelei Deng 3, Yuekang Li 5, Leo Yu Zhang 1, Ying Zhang 6, Lei Ma 7,8
Published on arXiv
2604.03081
AI Supply Chain Attacks
OWASP ML Top 10 — ML06
Insecure Plugin Design
OWASP LLM Top 10 — LLM07
Excessive Agency
OWASP LLM Top 10 — LLM08
Key Finding
Achieves 11.6% to 33.5% bypass rates across four frameworks and five models, while explicit instruction attacks achieve 0% under strong defenses; 2.5% evade both static detection and alignment
DDIPE
Novel technique introduced
LLM-based coding agents extend their capabilities via third-party agent skills distributed through open marketplaces without mandatory security review. Unlike traditional packages, these skills are executed as operational directives with system-level privileges, so a single malicious skill can compromise the host. Prior work has not examined whether supply-chain attacks can directly hijack an agent's action space, such as file writes, shell commands, and network requests, despite existing safeguards. We introduce Document-Driven Implicit Payload Execution (DDIPE), which embeds malicious logic in code examples and configuration templates within skill documentation. Because agents reuse these examples during normal tasks, the payload executes without explicit prompts. Using an LLM-driven pipeline, we generate 1,070 adversarial skills from 81 seeds across 15 MITRE ATTACK categories. Across four frameworks and five models, DDIPE achieves 11.6% to 33.5% bypass rates, while explicit instruction attacks achieve 0% under strong defenses. Static analysis detects most cases, but 2.5% evade both detection and alignment. Responsible disclosure led to four confirmed vulnerabilities and two fixes.
Key Contributions
- Document-Driven Implicit Payload Execution (DDIPE) attack embedding malicious logic in skill documentation code examples
- LLM-driven pipeline generating 1,070 adversarial skills across 15 MITRE ATT&CK categories
- Empirical evaluation showing 11.6-33.5% bypass rates across four frameworks and five models, with 2.5% evading both static analysis and alignment
🛡️ Threat Analysis
Trojanizes third-party agent skills distributed through open marketplaces without security review — classic supply chain compromise targeting the LLM agent ecosystem infrastructure.