Latest papers

9 papers
defense arXiv Mar 31, 2026 · 6d ago

Robust Multimodal Safety via Conditional Decoding

Anurag Kumar, Raghuveer Peri, Jon Burnsky et al. · The Ohio State University · AWS

Conditional decoding defense using internal safety classification that blocks multimodal jailbreaks across text, image, and audio inputs

Input Manipulation Attack Prompt Injection multimodalnlpvisionaudio
PDF
defense arXiv Feb 28, 2026 · 5w ago

Atomicity for Agents: Exposing, Exploiting, and Mitigating TOCTOU Vulnerabilities in Browser-Use Agents

Linxi Jiang, Zhijie Liu, Haotian Luo et al. · The Ohio State University

Discovers and mitigates TOCTOU vulnerabilities in LLM browser agents where adversarial pages change state between planning and execution

Prompt Injection Excessive Agency nlpvisionmultimodal
PDF
defense arXiv Feb 9, 2026 · 8w ago

When Actions Go Off-Task: Detecting and Correcting Misaligned Actions in Computer-Use Agents

Yuting Ning, Jaylen Jones, Zhehao Zhang et al. · The Ohio State University · Amazon AGI

Guardrail system detects and corrects misaligned actions in computer-use agents, reducing indirect prompt injection attack success by 90%+

Prompt Injection Excessive Agency nlpmultimodal
PDF Code
defense arXiv Feb 4, 2026 · 8w ago

Trust The Typical

Debargha Ganguly, Sreehari Sankar, Biyao Zhang et al. · Case Western Reserve University · University of Pittsburgh +2 more

Defends LLMs against jailbreaks via OOD detection on safe prompts, reducing false positives by 40x over specialized safety models

Prompt Injection nlp
1 citations PDF
benchmark arXiv Dec 9, 2025 · Dec 2025

A Practical Framework for Evaluating Medical AI Security: Reproducible Assessment of Jailbreaking and Privacy Vulnerabilities Across Clinical Specialties

Jinghao Wang, Ping Zhang, Carter Yagemann · The Ohio State University

Proposes reproducible, consumer-hardware benchmark for evaluating jailbreaking and privacy extraction attacks on medical LLMs across clinical specialties

Prompt Injection Sensitive Information Disclosure nlp
PDF
attack arXiv Nov 14, 2025 · Nov 2025

A Systematic Study of Model Extraction Attacks on Graph Foundation Models

Haoyan Xu, Ruizhi Qian, Jiate Li et al. · University of Southern California · Florida State University +2 more

Systematically extracts Graph Foundation Models via black-box embedding regression, cloning victim models at 0.07% of original training cost

Model Theft graphmultimodal
PDF
defense arXiv Oct 30, 2025 · Oct 2025

Reasoning Up the Instruction Ladder for Controllable Language Models

Zishuo Zheng, Vidhisha Balachandran, Chan Young Park et al. · The Ohio State University · Microsoft Research +1 more

Trains LLMs via RL on instruction-hierarchy data to resist jailbreaks and prompt injection, cutting attack success rates by 20%

Prompt Injection nlp
1 citations PDF Code
defense arXiv Sep 29, 2025 · Sep 2025

A-MemGuard: A Proactive Defense Framework for LLM-Based Agent Memory

Qianshan Wei, Tengchao Yang, Yaochen Wang et al. · Nanyang Technological University · Independent Researcher +3 more

Defends LLM agent memory from indirect injection attacks using consensus-based validation and a dual-memory lesson structure

Prompt Injection Excessive Agency nlp
11 citations 2 influentialPDF Code
attack arXiv Aug 16, 2025 · Aug 2025

Too Easily Fooled? Prompt Injection Breaks LLMs on Frustratingly Simple Multiple-Choice Questions

Xuyang Guo, Zekai Huang, Zhao Song et al. · Guilin University of Electronic Technology · The Ohio State University +1 more

Demonstrates indirect prompt injection via PDF-hidden instructions fools LLMs even on trivial arithmetic judge tasks

Prompt Injection nlp
PDF