Xiaojun Jia

Papers in Database (8)

attack arXiv Feb 15, 2026 · 7w ago

SkillJect: Automating Stealthy Skill-Based Prompt Injection for Coding Agents with Trace-Driven Closed-Loop Refinement

Xiaojun Jia, Jie Liao, Simeng Qin et al. · Nanyang Technological University · Chongqing University +4 more

Automated framework crafts stealthy skill-based prompt injections against LLM coding agents using closed-loop refinement agents

Prompt Injection Insecure Plugin Design nlp
PDF
attack arXiv Aug 7, 2025 · Aug 2025

PhysPatch: A Physically Realizable and Transferable Adversarial Patch Attack for Multimodal Large Language Models-based Autonomous Driving Systems

Qi Guo, Xiaojun Jia, Shanmin Pang et al. · Xi’an Jiaotong University · A*STAR +4 more

Physical adversarial patch attack on MLLM-based autonomous driving using SVD alignment and semantic mask optimization to steer perception and planning outputs

Input Manipulation Attack Prompt Injection visionmultimodal
PDF
attack arXiv Aug 6, 2025 · Aug 2025

The Emotional Baby Is Truly Deadly: Does your Multimodal Large Reasoning Model Have Emotional Flattery towards Humans?

Yuan Xun, Xiaojun Jia, Xinwei Liu et al. · Chinese Academy of Sciences · University of Chinese Academy of Sciences +1 more

EmoAgent jailbreaks multimodal reasoning models by using exaggerated emotional prompts to override safety protocols during deep-thinking stages

Prompt Injection multimodalnlp
PDF
defense arXiv Aug 5, 2025 · Aug 2025

GeoShield: Safeguarding Geolocation Privacy from Vision-Language Models via Adversarial Perturbations

Xinwei Liu, Xiaojun Jia, Yuan Xun et al. · Chinese Academy of Sciences · University of Chinese Academy of Sciences +3 more

Defends geolocation privacy against VLMs by applying adversarial image perturbations with feature disentanglement and scale-adaptive optimization

Input Manipulation Attack Prompt Injection visionmultimodal
PDF Code
attack The Fourteenth International C... Feb 28, 2026 · 5w ago

MIDAS: Multi-Image Dispersion and Semantic Reconstruction for Jailbreaking MLLMs

Yilian Liu, Xiaojun Jia, Guoshun Nan et al. · Beijing University of Posts and Telecommunications · Nanyang Technological University +1 more

Jailbreaks MLLMs by dispersing harmful semantics across multiple images, forcing cross-image reasoning that defeats safety alignment

Prompt Injection visionnlpmultimodal
PDF Code
defense arXiv Sep 18, 2025 · Sep 2025

LLM Jailbreak Detection for (Almost) Free!

Guorui Chen, Yifan Xia, Xiaojun Jia et al. · Wuhan University · Nanyang Technological University +1 more

Detects LLM jailbreaks near-free by comparing first-token confidence distributions between jailbreak and benign prompts

Prompt Injection nlp
PDF Code
defense arXiv Jan 7, 2025 · Jan 2025

PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models

Lingzhi Yuan, Xinfeng Li, Chejian Xu et al. · University of Maryland · Nanyang Technological University +2 more

Defends text-to-image models against NSFW prompt misuse via optimized safety soft prompts mimicking LLM system prompts

Prompt Injection visiongenerative
PDF
attack arXiv Feb 26, 2026 · 5w ago

Obscure but Effective: Classical Chinese Jailbreak Prompt Optimization via Bio-Inspired Search

Xun Huang, Simeng Qin, Xiaoshuang Jia et al. · Nanyang Technological University · BraneMatrix AI +7 more

Bio-inspired optimization generates classical Chinese jailbreak prompts that defeat modern-language safety guardrails in black-box LLMs

Prompt Injection nlp
PDF