Latest papers

2 papers
defense arXiv Jan 30, 2026 · 9w ago

FraudShield: Knowledge Graph Empowered Defense for LLMs against Fraud Attacks

Naen Xu, Jinghuai Zhang, Ping He et al. · Zhejiang University · University of California +1 more

Knowledge graph defense framework that detects fraud tactics in LLM inputs and augments prompts with evidence to resist manipulation

Prompt Injection nlp
PDF
attack arXiv Jan 9, 2026 · 12w ago

Knowledge-Driven Multi-Turn Jailbreaking on Large Language Models

Songze Li, Ruishi He, Xiaojun Jia et al. · Southeast University · Nanyang Technological University +1 more

Proposes Mastermind, a hierarchical multi-agent jailbreak framework that autonomously learns and adapts attack strategies across multi-turn LLM conversations

Prompt Injection nlp
1 citations PDF