Latest papers

2 papers
attack arXiv Mar 27, 2026 · 10d ago

H-Node Attack and Defense in Large Language Models

Eric Yocam, Varghese Vaidyan, Yong Wang · California Polytechnic State University · Dakota State University +1 more

Mechanistic attack amplifying hallucination nodes in LLM hidden states, with adaptive defense canceling excess activations at inference

Input Manipulation Attack Prompt Injection nlp
PDF
benchmark arXiv Dec 17, 2025 · Dec 2025

Quantifying Return on Security Controls in LLM Systems

Richard Helder Moulton, Austin O'Brien, John D. Hastings · Dakota State University

Quantifies financial return on LLM security controls via Monte Carlo simulation, showing ABAC cuts expected loss 94% while NeMo Guardrails barely helps

Prompt Injection Sensitive Information Disclosure nlp
PDF