survey 2026

Agentic AI as a Cybersecurity Attack Surface: Threats, Exploits, and Defenses in Runtime Supply Chains

Xiaochong Jiang 1, Shiqi Yang 2, Wenting Yang 3, Yichen Liu 3, Cheng Ji 4

0 citations · 44 references · arXiv (Cornell University)

α

Published on arXiv

2602.19555

Prompt Injection

OWASP LLM Top 10 — LLM01

Insecure Plugin Design

OWASP LLM Top 10 — LLM07

Excessive Agency

OWASP LLM Top 10 — LLM08

Key Finding

Identifies the Viral Agent Loop as a structurally novel threat enabling self-propagating AI worms through cyclic agentic execution, distinct from classical prompt injection or code-level exploits.

Viral Agent Loop

Novel technique introduced


Agentic systems built on large language models (LLMs) extend beyond text generation to autonomously retrieve information and invoke tools. This runtime execution model shifts the attack surface from build-time artifacts to inference-time dependencies, exposing agents to manipulation through untrusted data and probabilistic capability resolution. While prior work has focused on model-level vulnerabilities, security risks emerging from cyclic and interdependent runtime behavior remain fragmented. We systematize these risks within a unified runtime framework, categorizing threats into data supply chain attacks (transient context injection and persistent memory poisoning) and tool supply chain attacks (discovery, implementation, and invocation). We further identify the Viral Agent Loop, in which agents act as vectors for self-propagating generative worms without exploiting code-level flaws. Finally, we advocate a Zero-Trust Runtime Architecture that treats context as untrusted control flow and constrains tool execution through cryptographic provenance rather than semantic inference.


Key Contributions

  • Unified runtime supply chain framework categorizing agentic LLM threats into data supply chain attacks (context injection, memory poisoning) and tool supply chain attacks (discovery, implementation, invocation)
  • Introduction of the Viral Agent Loop — a self-propagating generative worm pattern where agent outputs re-enter as tainted context, enabling persistent compromise without code-level exploits
  • Proposal of a Zero-Trust Runtime Architecture using cryptographic provenance for tool execution rather than relying on semantic inference

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
inference_timeblack_box
Applications
llm agentstool-integrated llmsautonomous ai agents