Daeseon Choi

h-index: 3 48 citations 27 papers (total)

Papers in Database (2)

attack arXiv Oct 31, 2025 · Oct 2025

Self-HarmLLM: Can Large Language Model Harm Itself?

Heehwan Kim, Sungjune Park, Daeseon Choi · Soongsil University

Novel jailbreak attack where an LLM generates obfuscated harmful queries that bypass its own guardrails when re-entered in a new session

Prompt Injection nlp
PDF
defense arXiv Jan 13, 2026 · 11w ago

STAR: Detecting Inference-time Backdoors in LLM Reasoning via State-Transition Amplification Ratio

Seong-Gyu Park, Sohee Park, Jisu Lee et al. · Soongsil University

Detects inference-time backdoor triggers in LLM Chain-of-Thought reasoning via output probability shift analysis, achieving AUROC ≈ 1.0

Model Poisoning Prompt Injection nlp
PDF