Sungjune Park

h-index: 1 5 citations 5 papers (total)

Papers in Database (1)

attack arXiv Oct 31, 2025 · Oct 2025

Self-HarmLLM: Can Large Language Model Harm Itself?

Heehwan Kim, Sungjune Park, Daeseon Choi · Soongsil University

Novel jailbreak attack where an LLM generates obfuscated harmful queries that bypass its own guardrails when re-entered in a new session

Prompt Injection nlp
PDF