Yanjun Qi

h-index: 2 8 citations 4 papers (total)

Papers in Database (2)

attack arXiv Sep 30, 2025 · Sep 2025

STAC: When Innocent Tools Form Dangerous Chains to Jailbreak LLM Agents

Jing-Jing Li, Jianfeng He, Chao Shang et al. · AWS AI Labs · UC Berkeley

Multi-turn attack chains innocuous tool calls on LLM agents to achieve harmful goals, exceeding 90% ASR on GPT-4.1

Insecure Plugin Design Prompt Injection nlp
4 citations PDF Code
attack arXiv Oct 4, 2025 · Oct 2025

Cross-Modal Content Optimization for Steering Web Agent Preferences

Tanqiu Jiang, Min Bai, Nikolaos Pappas et al. · Stony Brook University · AWS AI Labs

Black-box attack jointly optimizes adversarial image perturbations and text to steer VLM web agent selection preferences

Input Manipulation Attack Prompt Injection visionnlpmultimodal
PDF