Yuxiao Li

h-index: 2 78 citations 7 papers (total)

Papers in Database (1)

attack arXiv Nov 8, 2025 · Nov 2025

Injecting Falsehoods: Adversarial Man-in-the-Middle Attacks Undermining Factual Recall in LLMs

Alina Fastowski, Bardh Prenkaj, Yuxiao Li et al. · Technical University of Munich

Proposes Xmera MitM framework injecting false factual contexts into LLM prompts, achieving 85% attack success, detected by uncertainty-based classifiers

Prompt Injection nlp
PDF Code