benchmark 2025

An Empirical Study on the Security Vulnerabilities of GPTs

Tong Wu , Weibin Wu , Zibin Zheng

0 citations · 71 references · arXiv

α

Published on arXiv

2512.00136

Prompt Injection

OWASP LLM Top 10 — LLM01

Sensitive Information Disclosure

OWASP LLM Top 10 — LLM06

Insecure Plugin Design

OWASP LLM Top 10 — LLM07

Key Finding

Top-ranked GPTs across all major categories are systematically vulnerable to prompt injection-based system prompt extraction and tool misuse attacks due to shared architectural patterns in the GPT framework.


Equipped with various tools and knowledge, GPTs, one kind of customized AI agents based on OpenAI's large language models, have illustrated great potential in many fields, such as writing, research, and programming. Today, the number of GPTs has reached three millions, with the range of specific expert domains becoming increasingly diverse. However, given the consistent framework shared among these LLM agent applications, systemic security vulnerabilities may exist and remain underexplored. To fill this gap, we present an empirical study on the security vulnerabilities of GPTs. Building upon prior research on LLM security, we first adopt a platform-user perspective to conduct a comprehensive attack surface analysis across different system components. Then, we design a systematic and multidimensional attack suite with the explicit objectives of information leakage and tool misuse based on the attack surface analysis, thereby concretely demonstrating the security vulnerabilities that various components of GPT-based systems face. Finally, we accordingly propose defense mechanisms to address the aforementioned security vulnerabilities. By increasing the awareness of these vulnerabilities and offering critical insights into their implications, this study seeks to facilitate the secure and responsible application of GPTs while contributing to developing robust defense mechanisms that protect users and systems against malicious attacks.


Key Contributions

  • Comprehensive attack surface analysis of GPT-based agents across four core components: expert prompt, chat history, tools, and knowledge
  • Systematic, multidimensional attack suite targeting information leakage and tool misuse across the top-ranked GPTs in the OpenAI GPT Store
  • Defense mechanism proposals addressing identified vulnerabilities across each system component

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
black_boxinference_time
Datasets
GPT Store top-ranked GPTs
Applications
llm agentsai assistantsgpt store applications