The Laminar Flow Hypothesis: Detecting Jailbreaks via Semantic Turbulence in Large Language Models
Published on arXiv
2512.13741
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Variance of layer-wise cosine velocity distinguishes jailbreak prompts from benign inputs with a 75.4% turbulence increase in RLHF-aligned Qwen2-1.5B (p < 0.001), enabling lightweight real-time detection without auxiliary classifiers.
Semantic Turbulence (Laminar Flow Hypothesis)
Novel technique introduced
As Large Language Models (LLMs) become ubiquitous, the challenge of securing them against adversarial "jailbreaking" attacks has intensified. Current defense strategies often rely on computationally expensive external classifiers or brittle lexical filters, overlooking the intrinsic dynamics of the model's reasoning process. In this work, the Laminar Flow Hypothesis is introduced, which posits that benign inputs induce smooth, gradual transitions in an LLM's high-dimensional latent space, whereas adversarial prompts trigger chaotic, high-variance trajectories - termed Semantic Turbulence - resulting from the internal conflict between safety alignment and instruction-following objectives. This phenomenon is formalized through a novel, zero-shot metric: the variance of layer-wise cosine velocity. Experimental evaluation across diverse small language models reveals a striking diagnostic capability. The RLHF-aligned Qwen2-1.5B exhibits a statistically significant 75.4% increase in turbulence under attack (p less than 0.001), validating the hypothesis of internal conflict. Conversely, Gemma-2B displays a 22.0% decrease in turbulence, characterizing a distinct, low-entropy "reflex-based" refusal mechanism. These findings demonstrate that Semantic Turbulence serves not only as a lightweight, real-time jailbreak detector but also as a non-invasive diagnostic tool for categorizing the underlying safety architecture of black-box models.
Key Contributions
- Introduces the Laminar Flow Hypothesis and formalizes Semantic Turbulence (variance of layer-wise cosine velocity) as a zero-shot, training-free jailbreak detection metric
- Empirically validates that RLHF-aligned models (Qwen2-1.5B) exhibit a statistically significant 75.4% turbulence increase under jailbreak attacks, while Gemma-2B shows a 22.0% decrease consistent with a 'reflex-based' refusal mechanism
- Demonstrates Semantic Turbulence as a non-invasive diagnostic tool for categorizing underlying LLM safety architectures (RLHF conflict-based vs. SFT reflex-based) without access to training data