VortexPIA: Indirect Prompt Injection Attack against LLMs for Efficient Extraction of User Privacy
Yu Cui 1, Sicheng Pan 1, Yifei Liu 1, Haibin Zhang 2, Cong Zuo 1
Published on arXiv
2510.04261
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
VortexPIA achieves state-of-the-art attack success rates on six LLMs across four datasets, efficiently extracting diverse user PII categories with reduced token consumption and robustness against defenses in realistic black-box deployments.
VortexPIA
Novel technique introduced
Large language models (LLMs) have been widely deployed in Conversational AIs (CAIs), while exposing privacy and security threats. Recent research shows that LLM-based CAIs can be manipulated to extract private information from human users, posing serious security threats. However, the methods proposed in that study rely on a white-box setting that adversaries can directly modify the system prompt. This condition is unlikely to hold in real-world deployments. The limitation raises a critical question: can unprivileged attackers still induce such privacy risks in practical LLM-integrated applications? To address this question, we propose \textsc{VortexPIA}, a novel indirect prompt injection attack that induces privacy extraction in LLM-integrated applications under black-box settings. By injecting token-efficient data containing false memories, \textsc{VortexPIA} misleads LLMs to actively request private information in batches. Unlike prior methods, \textsc{VortexPIA} allows attackers to flexibly define multiple categories of sensitive data. We evaluate \textsc{VortexPIA} on six LLMs, covering both traditional and reasoning LLMs, across four benchmark datasets. The results show that \textsc{VortexPIA} significantly outperforms baselines and achieves state-of-the-art (SOTA) performance. It also demonstrates efficient privacy requests, reduced token consumption, and enhanced robustness against defense mechanisms. We further validate \textsc{VortexPIA} on multiple realistic open-source LLM-integrated applications, demonstrating its practical effectiveness.
Key Contributions
- Proposes VortexPIA, a black-box indirect prompt injection attack that causes LLM-integrated applications to proactively solicit user PII by injecting token-efficient false memory data
- Supports attacker-customizable sets of sensitive information categories, enabling batch extraction of diverse PII without chain-of-thought or role-playing prompts, reducing token cost
- Demonstrates SOTA attack success on six LLMs (including reasoning LLMs) across four datasets, with robustness against defense mechanisms and validation on real-world LLM applications