Arjun Neekhra

h-index: 0 0 citations 2 papers (total)

Papers in Database (2)

benchmark arXiv Dec 11, 2025 · Dec 2025

How to Trick Your AI TA: A Systematic Study of Academic Jailbreaking in LLM Code Evaluation

Devanshu Sahoo, Vasudev Majhi, Arjun Neekhra et al.

Benchmarks 20+ jailbreak strategies against LLM-based academic code graders, achieving up to 97% success with persuasive and role-play attacks

Prompt Injection nlp
PDF
attack arXiv Jan 29, 2026 · 9w ago

The Compliance Paradox: Semantic-Instruction Decoupling in Automated Academic Code Evaluation

Devanshu Sahoo, Manish Prasad, Vasudev Majhi et al. · BITS Pilani · Trustwise +1 more

Embeds adversarial directives in AST comment nodes to hijack LLM-based code graders, achieving >95% manipulation success across 9 SOTA models

Prompt Injection nlp
PDF