Aiden Gabriel

h-index: 0 0 citations 2 papers (total)

Papers in Database (1)

defense arXiv Feb 19, 2026 · 6w ago

Fail-Closed Alignment for Large Language Models

Zachary Coalson, Beth Sohler, Aiden Gabriel et al. · Oregon State University

Defends LLMs against jailbreaks by training multiple independent refusal pathways that attackers cannot simultaneously suppress

Prompt Injection nlp
PDF