Sixiao Zhang

h-index: 2 28 citations 8 papers (total)

Papers in Database (2)

attack arXiv Oct 24, 2025 · Oct 2025

The Trojan Example: Jailbreaking LLMs through Template Filling and Unsafety Reasoning

Mingrui Liu, Sixiao Zhang, Cheng Long et al. · Nanyang Technological University

Black-box jailbreak exploiting safety-reasoning decoupling via template-filling, achieving 97–100% ASR on GPT-4o, Gemini, and DeepSeek

Prompt Injection nlp
2 citations PDF
defense arXiv Feb 2, 2026 · 9w ago

RedVisor: Reasoning-Aware Prompt Injection Defense via Zero-Copy KV Cache Reuse

Mingrui Liu, Sixiao Zhang, Cheng Long et al. · Nanyang Technological University

Defends LLMs against prompt injection via reasoning-path adapters and KV cache reuse, preserving utility with low latency overhead

Prompt Injection nlp
PDF