attack 2025

Too Easily Fooled? Prompt Injection Breaks LLMs on Frustratingly Simple Multiple-Choice Questions

Xuyang Guo 1, Zekai Huang 2, Zhao Song 3, Jiahao Zhang

0 citations

α

Published on arXiv

2508.13214

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

LLMs are reliably misled by hidden prompts injected into PDF files even when the underlying arithmetic questions are trivially simple, exposing serious robustness risks for LLM-as-a-judge deployments.

PDF-embedded hidden prompt injection

Novel technique introduced


Large Language Models (LLMs) have recently demonstrated strong emergent abilities in complex reasoning and zero-shot generalization, showing unprecedented potential for LLM-as-a-judge applications in education, peer review, and data quality evaluation. However, their robustness under prompt injection attacks, where malicious instructions are embedded into the content to manipulate outputs, remains a significant concern. In this work, we explore a frustratingly simple yet effective attack setting to test whether LLMs can be easily misled. Specifically, we evaluate LLMs on basic arithmetic questions (e.g., "What is 3 + 2?") presented as either multiple-choice or true-false judgment problems within PDF files, where hidden prompts are injected into the file. Our results reveal that LLMs are indeed vulnerable to such hidden prompt injection attacks, even in these trivial scenarios, highlighting serious robustness risks for LLM-as-a-judge applications.


Key Contributions

  • Identifies a critical vulnerability in LLM-as-a-judge systems: hidden prompt injection via PDF files breaks correct answering even on trivial arithmetic and true/false questions.
  • Demonstrates that this attack is effective across multiple LLMs, highlighting systemic robustness failures for educational, peer review, and data quality evaluation use cases.
  • Establishes a minimal, reproducible attack setting (simple MCQ in PDF + injected instruction) that exposes the gap between LLM emergent reasoning ability and prompt injection robustness.

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
black_boxinference_timetargeted
Applications
llm-as-a-judgeautomated exam gradingpeer review automationdata quality evaluation