Latest papers

4 papers
benchmark arXiv Feb 7, 2026 · 8w ago

Aegis: Towards Governance, Integrity, and Security of AI Voice Agents

Xiang Li, Pin-Yu Chen, Wenqi Wei · Fordham University · IBM Research

Red-teaming framework exposing behavioral vulnerabilities in AI voice agents via adversarial speech scenarios across banking, IT support, and logistics

Prompt Injection Excessive Agency audiomultimodalnlp
PDF
defense Quantum Machine Intelligence Jan 26, 2026 · 10w ago

Differentiable Architecture Search for Adversarially Robust Quantum Computer Vision

Mohamed Afane, Quanjiang Long, Haoting Shen et al. · Fordham University · Zhejiang University +2 more

Defends quantum neural networks against adversarial attacks via differentiable architecture search with trainable classical noise preprocessing

Input Manipulation Attack vision
PDF
defense BigData Congress Dec 10, 2025 · Dec 2025

SCOUT: A Defense Against Data Poisoning Attacks in Fine-Tuned Language Models

Mohamed Afane, Abhishek Satyam, Ke Chen et al. · Fordham University · Zhejiang University +2 more

SCOUT uses token-level saliency analysis to detect contextually-blended backdoor triggers in fine-tuned NLP models, including novel domain-specific attacks.

Model Poisoning Data Poisoning Attack nlp
PDF Code
attack arXiv Nov 20, 2025 · Nov 2025

An Image Is Worth Ten Thousand Words: Verbose-Text Induction Attacks on VLMs

Zhi Luo, Zenghui Yuan, Wenqi Wei et al. · Huazhong University of Science and Technology · Fordham University +1 more

Adversarial image perturbations force VLMs to generate verbose outputs via RL-optimized prompt embeddings, causing resource exhaustion DoS

Input Manipulation Attack Model Denial of Service visionmultimodalnlp
PDF