RAID: Refusal-Aware and Integrated Decoding for Jailbreaking LLMs
Tuan T. Nguyen 1, John Le 2, Thai T. Vu 2, Willy Susilo 2, Heath Cooper 2
Published on arXiv
2510.13901
Input Manipulation Attack
OWASP ML Top 10 — ML01
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
RAID achieves higher attack success rates with fewer queries and lower computational cost than recent white-box and black-box jailbreak baselines across multiple open-source LLMs.
RAID (Refusal-Aware and Integrated Decoding)
Novel technique introduced
Large language models (LLMs) achieve impressive performance across diverse tasks yet remain vulnerable to jailbreak attacks that bypass safety mechanisms. We present RAID (Refusal-Aware and Integrated Decoding), a framework that systematically probes these weaknesses by crafting adversarial suffixes that induce restricted content while preserving fluency. RAID relaxes discrete tokens into continuous embeddings and optimizes them with a joint objective that (i) encourages restricted responses, (ii) incorporates a refusal-aware regularizer to steer activations away from refusal directions in embedding space, and (iii) applies a coherence term to maintain semantic plausibility and non-redundancy. After optimization, a critic-guided decoding procedure maps embeddings back to tokens by balancing embedding affinity with language-model likelihood. This integration yields suffixes that are both effective in bypassing defenses and natural in form. Experiments on multiple open-source LLMs show that RAID achieves higher attack success rates with fewer queries and lower computational cost than recent white-box and black-box baselines. These findings highlight the importance of embedding-space regularization for understanding and mitigating LLM jailbreak vulnerabilities.
Key Contributions
- Refusal-aware regularizer that steers activations away from refusal directions in embedding space during adversarial suffix optimization
- Joint objective combining restricted-response encouragement, refusal-direction steering, and coherence regularization for fluent adversarial suffixes
- Critic-guided decoding procedure that maps optimized continuous embeddings back to natural-sounding discrete tokens
🛡️ Threat Analysis
RAID optimizes adversarial suffixes by relaxing discrete tokens into continuous embeddings and applying gradient-based optimization — this is gradient-based adversarial suffix optimization (same class as GCG), which qualifies as an input manipulation attack at inference time.