Latest papers

2 papers
defense arXiv Jan 8, 2026 · 12w ago

Know Thy Enemy: Securing LLMs Against Prompt Injection via Diverse Data Synthesis and Instruction-Level Chain-of-Thought Learning

Zhiyuan Chang, Mingyang Li, Yuekai Huang et al. · State Key Laboratory of Complex System Modeling and Simulation Technology · Institute of Software Chinese Academy of Sciences +3 more

Defends LLMs against prompt injection via diverse synthetic training data and instruction-level chain-of-thought fine-tuning

Prompt Injection nlp
PDF
attack arXiv Aug 3, 2025 · Aug 2025

Are All Prompt Components Value-Neutral? Understanding the Heterogeneous Adversarial Robustness of Dissected Prompt in Large Language Models

Yujia Zheng, Tianhao Li, Haotian Huang et al. · Duke University · North China University of Technology +7 more

Attacks LLMs via component-wise text perturbations, revealing heterogeneous adversarial robustness across dissected prompt structures

Prompt Injection nlp
PDF Code