defense 2025

Invasive Context Engineering to Control Large Language Models

Thomas Rivasseau

0 citations · 38 references · arXiv

α

Published on arXiv

2512.03001

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

Argues theoretically that periodically injecting control sentences into LLM context maintains safety guarantees more scalably than training-based approaches, which require exponentially more data as context length grows.

Invasive Context Engineering (ICE)

Novel technique introduced


Current research on operator control of Large Language Models improves model robustness against adversarial attacks and misbehavior by training on preference examples, prompting, and input/output filtering. Despite good results, LLMs remain susceptible to abuse, and jailbreak probability increases with context length. There is a need for robust LLM security guarantees in long-context situations. We propose control sentences inserted into the LLM context as invasive context engineering to partially solve the problem. We suggest this technique can be generalized to the Chain-of-Thought process to prevent scheming. Invasive Context Engineering does not rely on LLM training, avoiding data shortage pitfalls which arise in training models for long context situations.


Key Contributions

  • Formalizes the 'long-context problem': exponential training data requirements for alignment coverage (Ω(k^l)) and diminishing system prompt influence as context grows (lim s/l → 0).
  • Proposes Invasive Context Engineering (ICE) — training-free runtime insertion of control sentences into LLM contexts to reinforce safety guidelines throughout long conversations.
  • Suggests ICE can be generalized to Chain-of-Thought processes to counteract scheming behaviors in reasoning-capable frontier models.

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
inference_time
Applications
llm safetylong-context chatbotsreasoning models with chain-of-thought