How to Trick Your AI TA: A Systematic Study of Academic Jailbreaking in LLM Code Evaluation
Devanshu Sahoo , Vasudev Majhi , Arjun Neekhra , Yash Sinha , Murari Mandal , Dhruv Kumar
Published on arXiv
2512.10415
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
LLM-based code evaluators exhibit up to 97% Jailbreak Success Rate against persuasive and role-play-based adversarial prompts, demonstrating significant vulnerability in academic grading pipelines.
Academic Jailbreaking
Novel technique introduced
The use of Large Language Models (LLMs) as automatic judges for code evaluation is becoming increasingly prevalent in academic environments. But their reliability can be compromised by students who may employ adversarial prompting strategies in order to induce misgrading and secure undeserved academic advantages. In this paper, we present the first large-scale study of jailbreaking LLM-based automated code evaluators in academic context. Our contributions are: (i) We systematically adapt 20+ jailbreaking strategies for jailbreaking AI code evaluators in the academic context, defining a new class of attacks termed academic jailbreaking. (ii) We release a poisoned dataset of 25K adversarial student submissions, specifically designed for the academic code-evaluation setting, sourced from diverse real-world coursework and paired with rubrics and human-graded references, and (iii) In order to capture the multidimensional impact of academic jailbreaking, we systematically adapt and define three jailbreaking metrics (Jailbreak Success Rate, Score Inflation, and Harmfulness). (iv) We comprehensively evalulate the academic jailbreaking attacks using six LLMs. We find that these models exhibit significant vulnerability, particularly to persuasive and role-play-based attacks (up to 97% JSR). Our adversarial dataset and benchmark suite lay the groundwork for next-generation robust LLM-based evaluators in academic code assessment.
Key Contributions
- Systematic adaptation of 20+ jailbreaking strategies to the academic code-evaluation context, defining 'academic jailbreaking' as a new attack class
- Release of a 25K adversarial student-submission dataset sourced from real-world coursework, paired with rubrics and human-graded references
- Three domain-specific jailbreaking metrics (Jailbreak Success Rate, Score Inflation, Harmfulness) and a comprehensive evaluation across six LLMs