Hyunsik Na

h-index: 3 41 citations 18 papers (total)

Papers in Database (2)

benchmark arXiv Nov 18, 2025 · Nov 2025

Beyond Fixed and Dynamic Prompts: Embedded Jailbreak Templates for Advancing LLM Security

Hajun Kim, Hyunsik Na, Daeseon Choi · Soongsil University

Proposes Embedded Jailbreak Templates that naturally integrate harmful queries into existing prompt structures for more realistic LLM red-teaming benchmarks

Prompt Injection nlp
PDF
defense arXiv Jan 13, 2026 · 11w ago

STAR: Detecting Inference-time Backdoors in LLM Reasoning via State-Transition Amplification Ratio

Seong-Gyu Park, Sohee Park, Jisu Lee et al. · Soongsil University

Detects inference-time backdoor triggers in LLM Chain-of-Thought reasoning via output probability shift analysis, achieving AUROC ≈ 1.0

Model Poisoning Prompt Injection nlp
PDF