Is Reasoning Capability Enough for Safety in Long-Context Language Models?
Yu Fu 1, Haz Sameen Shahgir 1, Huanli Gong 2,3, Zhipeng Wei 3, N. Benjamin Erichson 3,4, Yue Dong 1
1 University of California, Riverside
2 University of California, Berkeley
3 International Computer Science University of California, Berkeley
Published on arXiv
2602.08874
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Stronger reasoning capability does not improve LLM safety under compositional attacks; safety alignment degrades with context length, but increasing inference-time compute reduces attack success rate by over 50 percentage points.
Compositional Reasoning Attacks
Novel technique introduced
Large language models (LLMs) increasingly combine long-context processing with advanced reasoning, enabling them to retrieve and synthesize information distributed across tens of thousands of tokens. A hypothesis is that stronger reasoning capability should improve safety by helping models recognize harmful intent even when it is not stated explicitly. We test this hypothesis in long-context settings where harmful intent is implicit and must be inferred through reasoning, and find that it does not hold. We introduce compositional reasoning attacks, a new threat model in which a harmful query is decomposed into incomplete fragments that scattered throughout a long context. The model is then prompted with a neutral reasoning query that induces retrieval and synthesis, causing the harmful intent to emerge only after composition. Evaluating 14 frontier LLMs on contexts up to 64k tokens, we uncover three findings: (1) models with stronger general reasoning capability are not more robust to compositional reasoning attacks, often assembling the intent yet failing to refuse; (2) safety alignment consistently degrades as context length increases; and (3) inference-time reasoning effort is a key mitigating factor: increasing inference-time compute reduces attack success by over 50 percentage points on GPT-oss-120b model. Together, these results suggest that safety does not automatically scale with reasoning capability, especially under long-context inference.
Key Contributions
- Introduces compositional reasoning attacks: a single-pass, long-context threat model where a harmful query is decomposed into incomplete fragments scattered throughout the context, with harmful intent emerging only after model synthesis
- Empirically demonstrates across 14 frontier LLMs that stronger general reasoning capability does not improve robustness to this attack, and that safety alignment consistently degrades as context length increases up to 64k tokens
- Identifies inference-time compute as a key mitigating factor, with increased reasoning effort reducing attack success by over 50 percentage points on GPT-oss-120b