Latest papers

2 papers
attack arXiv Feb 3, 2026 · 8w ago

Controlling Output Rankings in Generative Engines for LLM-based Search

Haibo Jin, Ruoxi Chen, Peiyan Zhang et al. · University of Illinois at Urbana-Champaign · Starc Institute +2 more

Injects crafted content into product pages to manipulate LLM-based search rankings with 91% promotion success rate

Input Manipulation Attack Prompt Injection nlp
PDF
tool arXiv Aug 28, 2025 · Aug 2025

GUARD: Guideline Upholding Test through Adaptive Role-play and Jailbreak Diagnostics for LLMs

Haibo Jin, Ruoxi Chen, Peiyan Zhang et al. · University of Illinois at Urbana-Champaign · Starc Institute +1 more

Automated LLM red-teaming tool translates government AI ethics guidelines into jailbreak diagnostics and compliance reports

Prompt Injection nlpmultimodal
PDF