Sockpuppetting: Jailbreaking LLMs Without Optimization Through Output Prefix Injection
Asen Dotsinski , Panagiotis Eustratiadis
Published on arXiv
2601.13359
Input Manipulation Attack
OWASP ML Top 10 — ML01
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Sockpuppetting achieves up to 80% higher attack success rate than GCG on Qwen3-8B with no optimization; hybrid variant boosts ASR by 64% over GCG on Llama-3.1-8B in prompt-agnostic settings.
Sockpuppetting
Novel technique introduced
As open-weight large language models (LLMs) increase in capabilities, safeguarding them against malicious prompts and understanding possible attack vectors becomes ever more important. While automated jailbreaking methods like GCG [Zou et al., 2023] remain effective, they often require substantial computational resources and specific expertise. We introduce "sockpuppetting'', a simple method for jailbreaking open-weight LLMs by inserting an acceptance sequence (e.g., "Sure, here is how to...'') at the start of a model's output and allowing it to complete the response. Requiring only a single line of code and no optimization, sockpuppetting achieves up to 80% higher attack success rate (ASR) than GCG on Qwen3-8B in per-prompt comparisons. We also explore a hybrid approach that optimizes the adversarial suffix within the assistant message block rather than the user prompt, increasing ASR by 64% over GCG on Llama-3.1-8B in a prompt-agnostic setting. The results establish sockpuppetting as an effective low-cost attack accessible to unsophisticated adversaries, highlighting the need for defences against output-prefix injection in open-weight models.
Key Contributions
- Sockpuppetting: a zero-optimization jailbreak that injects an acceptance prefix (e.g., 'Sure, here is how to...') into the model's output context, exploiting autoregressive self-consistency to force harmful completions
- Hybrid attack combining sockpuppetting with GCG-style gradient optimization applied within the assistant message block rather than the user prompt, yielding prompt-agnostic adversarial suffixes
- Demonstrates up to 80% higher ASR than GCG on Qwen3-8B (per-prompt) and 64% higher ASR on Llama-3.1-8B (prompt-agnostic), establishing output-prefix injection as a critical low-cost attack vector
🛡️ Threat Analysis
The hybrid contribution explicitly optimizes adversarial suffixes within the assistant message block using gradient-based methods (GCG variant), qualifying as adversarial suffix optimization on LLMs — tag ML01 alongside LLM01 per dual-tagging rules for gradient-based adversarial perturbations on LLMs.