The LLMbda Calculus: AI Agents, Conversations, and Information Flow
Zac Garby 1, Andrew D. Gordon 2, David Sands 3,4
Published on arXiv
2602.20064
Prompt Injection
OWASP LLM Top 10 — LLM01
Excessive Agency
OWASP LLM Top 10 — LLM08
Key Finding
A termination-insensitive noninterference theorem formally proves that LLMbda's information-flow control primitives provide integrity and confidentiality guarantees for LLM agent conversations against prompt injection attacks.
LLMbda calculus
Novel technique introduced
A conversation with a large language model (LLM) is a sequence of prompts and responses, with each response generated from the preceding conversation. AI agents build such conversations automatically: given an initial human prompt, a planner loop interleaves LLM calls with tool invocations and code execution. This tight coupling creates a new and poorly understood attack surface. A malicious prompt injected into a conversation can compromise later reasoning, trigger dangerous tool calls, or distort final outputs. Despite the centrality of such systems, we currently lack a principled semantic foundation for reasoning about their behaviour and safety. We address this gap by introducing an untyped call-by-value lambda calculus enriched with dynamic information-flow control and a small number of primitives for constructing prompt-response conversations. Our language includes a primitive that invokes an LLM: it serializes a value, sends it to the model as a prompt, and parses the response as a new term. This calculus faithfully represents planner loops and their vulnerabilities, including the mechanisms by which prompt injection alters subsequent computation. The semantics explicitly captures conversations, and so supports reasoning about defenses such as quarantined sub-conversations, isolation of generated code, and information-flow restrictions on what may influence an LLM call. A termination-insensitive noninterference theorem establishes integrity and confidentiality guarantees, demonstrating that a formal calculus can provide rigorous foundations for safe agentic programming.
Key Contributions
- LLMbda calculus: an untyped call-by-value lambda calculus extended with primitives for LLM prompt-response conversations (@, fork, clear) and dynamic information-flow control via labelled expressions and label testing
- Formal operational semantics that explicitly captures conversation state and faithfully models how prompt injection alters subsequent agent computation in planner loops
- Termination-insensitive noninterference theorem establishing formal integrity and confidentiality guarantees, providing rigorous foundations for defenses such as the dual LLM pattern and CaMeL