Latest papers

2 papers
benchmark arXiv Oct 24, 2025 · Oct 2025

Quantifying CBRN Risk in Frontier Models

Divyanshu Kumar, Nitin Aravind Birur, Tanay Baswa et al. · Enkrypt AI

Benchmarks 10 frontier LLMs against CBRN jailbreak prompts, finding Deep Inception attacks bypass safety filters 86% of the time versus 34% for direct requests

Prompt Injection nlp
2 citations PDF
attack arXiv Oct 23, 2025 · Oct 2025

Beyond Text: Multimodal Jailbreaking of Vision-Language and Audio Models through Perceptually Simple Transformations

Divyanshu Kumar, Shreyas Jena, Nitin Aravind Birur et al. · Enkrypt AI

Systematic multimodal jailbreak study shows simple image/audio transformations achieve 75–89% ASR on frontier VLMs with near-perfect text safety

Prompt Injection visionaudiomultimodalnlp
PDF