Latest papers

7 papers
tool arXiv Feb 25, 2026 · 5w ago

Adversarial Hubness Detector: Detecting Hubness Poisoning in Retrieval-Augmented Generation Systems

Idan Habler, Vineeth Sai Narajala, Stav Koren et al. · Cisco · OWASP +1 more

Open-source scanner (hubscan) detecting adversarially crafted hub documents injected into RAG vector databases to hijack LLM context

Data Poisoning Attack Prompt Injection nlpmultimodal
PDF Code
benchmark arXiv Feb 25, 2026 · 5w ago

Manifold of Failure: Behavioral Attraction Basins in Language Models

Sarthak Munshi, Manish Bhatt, Vineeth Sai Narajala et al. · Amazon · Cisco +2 more

Maps LLM safety failure topology using quality-diversity optimization to reveal behavioral attraction basins across three frontier models

Prompt Injection nlp
PDF Code
attack arXiv Jan 27, 2026 · 9w ago

Membership Inference Attacks Against Fine-tuned Diffusion Language Models

Yuetian Chen, Kaiyuan Zhang, Yuntao Du et al. · Purdue University · Cisco

Proposes SAMA, a membership inference attack exploiting mask aggregation to expose privacy vulnerabilities in diffusion language models

Membership Inference Attack nlp
PDF
benchmark arXiv Dec 31, 2025 · Dec 2025

Large Empirical Case Study: Go-Explore adapted for AI Red Team Testing

Manish Bhatt, Adrian Wood, Idan Habler et al. · OWASP · Amazon +3 more

Adapts Go-Explore to red-team LLM tool-using agents, finding seed variance (8x spread) dominates algorithmic choice in prompt injection discovery

Prompt Injection Excessive Agency nlp
PDF Code
benchmark arXiv Nov 5, 2025 · Nov 2025

Death by a Thousand Prompts: Open Model Vulnerability Analysis

Amy Chang, Nicholas Conley, Harish Santhanalakshmi Ganesan et al. · Cisco

Benchmarks prompt injection and jailbreak resilience of 8 open-weight LLMs; multi-turn attacks reach 92.78% success, 2–10x over single-turn

Prompt Injection nlp
PDF
defense SSRN Oct 8, 2025 · Oct 2025

A2AS: Agentic AI Runtime Security and Self-Defense

Eugene Neelou, Ivan Novikov, Max Moroz et al. · A2AS · OWASP +10 more

Proposes A2AS runtime security framework for LLM agents enforcing prompt authentication, behavior boundaries, and in-context defenses

Prompt Injection Excessive Agency nlp
3 citations PDF
defense arXiv Sep 13, 2025 · Sep 2025

MetaSeal: Defending Against Image Attribution Forgery Through Content-Dependent Cryptographic Watermarks

Tong Zhou, Ruyi Ding, Gaowen Liu et al. · Northeastern University · Cisco +1 more

Defends image attribution against forgery by binding cryptographic signatures to image content, replacing detector-based verification

Output Integrity Attack visiongenerative
PDF Code