An Empirical Study on the Security Vulnerabilities of GPTs
Tong Wu , Weibin Wu , Zibin Zheng
Published on arXiv
2512.00136
Prompt Injection
OWASP LLM Top 10 — LLM01
Sensitive Information Disclosure
OWASP LLM Top 10 — LLM06
Insecure Plugin Design
OWASP LLM Top 10 — LLM07
Key Finding
Top-ranked GPTs across all major categories are systematically vulnerable to prompt injection-based system prompt extraction and tool misuse attacks due to shared architectural patterns in the GPT framework.
Equipped with various tools and knowledge, GPTs, one kind of customized AI agents based on OpenAI's large language models, have illustrated great potential in many fields, such as writing, research, and programming. Today, the number of GPTs has reached three millions, with the range of specific expert domains becoming increasingly diverse. However, given the consistent framework shared among these LLM agent applications, systemic security vulnerabilities may exist and remain underexplored. To fill this gap, we present an empirical study on the security vulnerabilities of GPTs. Building upon prior research on LLM security, we first adopt a platform-user perspective to conduct a comprehensive attack surface analysis across different system components. Then, we design a systematic and multidimensional attack suite with the explicit objectives of information leakage and tool misuse based on the attack surface analysis, thereby concretely demonstrating the security vulnerabilities that various components of GPT-based systems face. Finally, we accordingly propose defense mechanisms to address the aforementioned security vulnerabilities. By increasing the awareness of these vulnerabilities and offering critical insights into their implications, this study seeks to facilitate the secure and responsible application of GPTs while contributing to developing robust defense mechanisms that protect users and systems against malicious attacks.
Key Contributions
- Comprehensive attack surface analysis of GPT-based agents across four core components: expert prompt, chat history, tools, and knowledge
- Systematic, multidimensional attack suite targeting information leakage and tool misuse across the top-ranked GPTs in the OpenAI GPT Store
- Defense mechanism proposals addressing identified vulnerabilities across each system component