Latest papers

2 papers
attack Applied Informatics Jan 20, 2026 · 10w ago

LLM Security and Safety: Insights from Homotopy-Inspired Prompt Obfuscation

Luis Lazo, Hamed Jelodar, Roozbeh Razavi-Far · University of New Brunswick

Jailbreaks LLMs for malicious code generation using homotopy-inspired prompt obfuscation across 15,732 engineered prompts

Prompt Injection nlp
PDF
defense arXiv Dec 23, 2025 · Dec 2025

Divided We Fall: Defending Against Adversarial Attacks via Soft-Gated Fractional Mixture-of-Experts with Randomized Adversarial Training

Mohammad Meymani, Roozbeh Razavi-Far · University of New Brunswick

Mixture-of-Experts defense with adversarial training outperforms state-of-the-art against white-box FGSM and PGD attacks on CIFAR-10 and SVHN

Input Manipulation Attack vision
PDF