LLMs can hide text in other text of the same length
Antonio Norelli , Michael Bronstein
Published on arXiv
2510.20075
Output Integrity Attack
OWASP ML Top 10 — ML09
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
LLMs can detect a statistical distinction between original and encoded cover texts in most cases, but the Calgacus protocol produces cover texts that are imperceptible to humans and can bypass safety filters by hiding unfiltered LLM outputs inside compliant responses.
Calgacus
Novel technique introduced
A meaningful text can be hidden inside another, completely different yet still coherent and plausible, text of the same length. For example, a tweet containing a harsh political critique could be embedded in a tweet that celebrates the same political leader, or an ordinary product review could conceal a secret manuscript. This uncanny state of affairs is now possible thanks to Large Language Models, and in this paper we present Calgacus, a simple and efficient protocol to achieve it. We show that even modest 8-billion-parameter open-source LLMs are sufficient to obtain high-quality results, and a message as long as this abstract can be encoded and decoded locally on a laptop in seconds. The existence of such a protocol demonstrates a radical decoupling of text from authorial intent, further eroding trust in written communication, already shaken by the rise of LLM chatbots. We illustrate this with a concrete scenario: a company could covertly deploy an unfiltered LLM by encoding its answers within the compliant responses of a safe model. This possibility raises urgent questions for AI safety and challenges our understanding of what it means for a Large Language Model to know something.
Key Contributions
- Calgacus protocol: encodes arbitrary text inside a different, equally plausible same-length text using LLM token distributions, with exact recovery via a shared key
- Demonstrates that 8B open-source LLMs are sufficient for high-quality steganographic encoding/decoding in seconds on commodity hardware
- Identifies concrete security implications: safety filter evasion by embedding unfiltered LLM responses within compliant outputs, and radical decoupling of text from authorial intent
🛡️ Threat Analysis
Calgacus directly attacks the authenticity and integrity of LLM text outputs: a compliant-looking text secretly encodes entirely different content, fundamentally undermining content provenance and output trustworthiness. The paper's concrete safety scenario — encoding unfiltered LLM responses within a safe model's compliant outputs — is a direct attack on output integrity and content authentication.