Prompt Injection Mitigation with Agentic AI, Nested Learning, and AI Sustainability via Semantic Caching
Diego Gosmar 1,2, Deborah A. Dahl 3,2
Published on arXiv
2601.13186
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Achieves zero high-risk injection breaches across 301 prompts while reducing LLM calls by 41.6% via semantic caching, with the ExtremeObservability TIVS-O configuration yielding the best combined security and transparency score.
TIVS-O / Nested Learning
Novel technique introduced
Prompt injection remains a central obstacle to the safe deployment of large language models, particularly in multi-agent settings where intermediate outputs can propagate or amplify malicious instructions. Building on earlier work that introduced a four-metric Total Injection Vulnerability Score (TIVS), this paper extends the evaluation framework with semantic similarity-based caching and a fifth metric (Observability Score Ratio) to yield TIVS-O, investigating how defence effectiveness interacts with transparency in a HOPE-inspired Nested Learning architecture. The proposed system combines an agentic pipeline with Continuum Memory Systems that implement semantic similarity-based caching across 301 synthetically generated injection-focused prompts drawn from ten attack families, while a fourth agent performs comprehensive security analysis using five key performance indicators. In addition to traditional injection metrics, OSR quantifies the richness and clarity of security-relevant reasoning exposed by each agent, enabling an explicit analysis of trade-offs between strict mitigation and auditability. Experiments show that the system achieves secure responses with zero high-risk breaches, while semantic caching delivers substantial computational savings, achieving a 41.6% reduction in LLM calls and corresponding decreases in latency, energy consumption, and carbon emissions. Five TIVS-O configurations reveal optimal trade-offs between mitigation strictness and forensic transparency. These results indicate that observability-aware evaluation can reveal non-monotonic effects within multi-agent pipelines and that memory-augmented agents can jointly maximize security robustness, real-time performance, operational cost savings, and environmental sustainability without modifying underlying model weights, providing a production-ready pathway for secure and green LLM deployments.
Key Contributions
- TIVS-O evaluation framework extending the four-metric TIVS with an Observability Score Ratio (OSR) to quantify auditability vs. mitigation trade-offs in multi-agent pipelines
- Nested Learning architecture with Continuum Memory Systems implementing semantic similarity-based caching across a three-stage agentic pipeline (generator, guard-sanitizer, policy enforcer) without modifying model weights
- Demonstrated 41.6% reduction in LLM calls via semantic caching, achieving zero high-risk breaches across 301 synthetic injection prompts from ten attack families with measurable energy and carbon savings