Latest papers

291 papers
defense arXiv Apr 1, 2026 · 5d ago

RAGShield: Provenance-Verified Defense-in-Depth Against Knowledge Base Poisoning in Government Retrieval-Augmented Generation Systems

KrishnaSaiReddy Patil

Defense-in-depth framework using cryptographic provenance verification to block knowledge base poisoning attacks in government RAG systems

Data Poisoning Attack Training Data Poisoning nlp
PDF
survey arXiv Apr 1, 2026 · 5d ago

Safety, Security, and Cognitive Risks in World Models

Manoj Parmar · SovereignAI Security Labs

Unified threat model for world model AI systems covering adversarial attacks, data poisoning, alignment risks, and cognitive security

Input Manipulation Attack Data Poisoning Attack Model Poisoning Prompt Injection Excessive Agency reinforcement-learningmultimodalvisionnlp
PDF
attack arXiv Apr 1, 2026 · 5d ago

Thinking Wrong in Silence: Backdoor Attacks on Continuous Latent Reasoning

Swapnil Parekh · Intuit

Backdoor attack on tokenless reasoning models that hijacks continuous latent trajectories via single embedding perturbations, achieving 99%+ success while evading all token-level defenses

Model Poisoning Data Poisoning Attack nlp
PDF
attack arXiv Mar 31, 2026 · 6d ago

Beyond Corner Patches: Semantics-Aware Backdoor Attack in Federated Learning

Kavindu Herath, Joshua Zhao, Saurabh Bagchi · Purdue University

Backdoor attack on federated learning using semantic triggers like sunglasses that evade robust aggregation defenses

Model Poisoning Data Poisoning Attack visionfederated-learning
PDF
defense arXiv Mar 30, 2026 · 7d ago

Mitigating Backdoor Attacks in Federated Learning Using PPA and MiniMax Game Theory

Osama Wehbi, Sarhad Arisdakessian, Omar Abdel Wahab et al. · Polytechnique Montréal · Institut national de la recherche scientifique +2 more

Defends federated learning against backdoor attacks using reputation systems, game theory, and statistical analysis to reduce attack success to 1-11%

Model Poisoning Data Poisoning Attack visionfederated-learning
PDF
defense arXiv Mar 30, 2026 · 7d ago

FedFG: Privacy-Preserving and Robust Federated Learning via Flow-Matching Generation

Ruiyang Wang, Rong Pan, Zhengan Yao · Sun Yat-Sen University

Federated learning defense using flow-matching generators to prevent gradient inversion and detect poisoning attacks simultaneously

Data Poisoning Attack Model Inversion Attack federated-learningvision
PDF Code
benchmark arXiv Mar 27, 2026 · 10d ago

Are LLM-Enhanced Graph Neural Networks Robust against Poisoning Attacks?

Yuhang Ma, Jie Wang, Zheng Yan · Xidian University · Hangzhou Institute of Technology

Benchmark evaluating LLM-enhanced GNNs against structural and textual poisoning attacks, finding them more robust than baseline embeddings

Data Poisoning Attack graphnlp
PDF Code
defense arXiv Mar 26, 2026 · 11d ago

Agentic Trust Coordination for Federated Learning through Adaptive Thresholding and Autonomous Decision Making in Sustainable and Resilient Industrial Networks

Paul Shepherd, Tasos Dagiuklas, Bugra Alkan et al. · London South Bank University · Instituto de Telecomunicações

Agentic control layer for federated learning that adaptively adjusts trust thresholds to defend against Byzantine and poisoning attacks

Data Poisoning Attack federated-learning
PDF
attack arXiv Mar 26, 2026 · 11d ago

PIDP-Attack: Combining Prompt Injection with Database Poisoning Attacks on Retrieval-Augmented Generation Systems

Haozhen Wang, Haoyue Liu, Jionghao Zhu et al. · The Chinese University of Hong Kong · Taobao and Tmall Group

Combines prompt injection with database poisoning to manipulate RAG system outputs for arbitrary queries without knowing them beforehand

Input Manipulation Attack Data Poisoning Attack Prompt Injection nlp
PDF
survey arXiv Mar 25, 2026 · 12d ago

AI Security in the Foundation Model Era: A Comprehensive Survey from a Unified Perspective

Zhenyi Wang, Siyu Luan · University of Central Florida · University of Copenhagen

Unified taxonomy of ML security threats organizing attacks into data-to-data, data-to-model, model-to-data, and model-to-model categories

Input Manipulation Attack Data Poisoning Attack Model Inversion Attack Membership Inference Attack Model Theft Output Integrity Attack Model Poisoning Prompt Injection Sensitive Information Disclosure visionnlpmultimodal
PDF
defense arXiv Mar 25, 2026 · 12d ago

DP^2-VL: Private Photo Dataset Protection by Data Poisoning for Vision-Language Models

Hongyi Miao, Jun Jia, Xincheng Wang et al. · Shandong University · Shanghai Jiao Tong University +4 more

Data poisoning defense that protects private photo datasets from VLM fine-tuning attacks that extract identity-affiliation relationships

Data Poisoning Attack Sensitive Information Disclosure visionnlpmultimodal
PDF
defense arXiv Mar 24, 2026 · 13d ago

ProGRank: Probe-Gradient Reranking to Defend Dense-Retriever RAG from Corpus Poisoning

Xiangyu Yin, Yi Qi, Chih-hong Cheng · Chalmers University of Technology · Carl von Ossietzky University of Oldenburg +1 more

Reranking defense for RAG that detects corpus-poisoned passages using gradient-based instability signals under perturbations

Data Poisoning Attack Prompt Injection nlp
PDF
defense arXiv Mar 24, 2026 · 13d ago

Byzantine-Robust and Differentially Private Federated Optimization under Weaker Assumptions

Rustem Islamov, Grigory Malinovsky, Alexander Gaponov et al. · University of Basel · KAUST +1 more

Byzantine-robust federated learning with differential privacy, proving convergence without bounded gradient assumptions using double momentum and clipping

Data Poisoning Attack federated-learning
PDF
attack arXiv Mar 24, 2026 · 13d ago

PoiCGAN: A Targeted Poisoning Based on Feature-Label Joint Perturbation in Federated Learning

Tao Liu, Jiguang Lv, Dapeng Man et al. · Harbin Engineering University

Targeted federated learning poisoning attack using CGAN-based sample generation achieving 84% higher success than baselines while evading detection

Data Poisoning Attack Model Poisoning visionfederated-learning
PDF
benchmark arXiv Mar 21, 2026 · 16d ago

Unveiling the Security Risks of Federated Learning in the Wild: From Research to Practice

Jiahao Chen, Zhiming Zhao, Yuwen Pu et al. · Zhejiang University · Chongqing University +1 more

Measurement study showing FL poisoning attacks are less effective in practice than research suggests due to heterogeneity and stability constraints

Data Poisoning Attack visionnlptabularfederated-learning
PDF Code
attack arXiv Mar 21, 2026 · 16d ago

Adversarial Attacks on Locally Private Graph Neural Networks

Matta Varun, Ajay Kumar Dhakar, Yuan Hong et al. · Indian Institute of Technology Kharagpur · University of Connecticut

Analyzes adversarial attacks on LDP-protected GNNs, exploring how privacy noise affects attack effectiveness and robustness

Input Manipulation Attack Data Poisoning Attack graph
PDF
attack arXiv Mar 20, 2026 · 17d ago

Graph-Aware Text-Only Backdoor Poisoning for Text-Attributed Graphs

Qi Luo, Minghui Xu, Dongxiao Yu et al. · Shandong University

Text-only backdoor attack on graph neural networks that poisons node text while preserving graph structure, achieving near-perfect attack success rates

Model Poisoning Data Poisoning Attack nlpgraph
PDF
defense arXiv Mar 20, 2026 · 17d ago

ARMOR: Adaptive Resilience Against Model Poisoning Attacks in Continual Federated Learning for Mobile Indoor Localization

Danish Gufran, Akhil Singampalli, Sudeep Pasricha · Colorado State University

Defends federated learning for indoor localization against model poisoning by predicting expected weight updates with state-space models

Data Poisoning Attack federated-learning
PDF
defense arXiv Mar 20, 2026 · 17d ago

Memory poisoning and secure multi-agent systems

Vicenç Torra, Maria Bras-Amorós · Umeå University · Universitat Politècnica de Catalunya

Defends LLM-based agents against memory poisoning attacks across semantic, episodic, and short-term memory using cryptographic techniques

Data Poisoning Attack Excessive Agency nlp
PDF
defense arXiv Mar 19, 2026 · 18d ago

FedAgain: A Trust-Based and Robust Federated Learning Strategy for an Automated Kidney Stone Identification in Ureteroscopy

Ivan Reyes-Amezcua, Francisco Lopez-Tiro, Clément Larose et al. · Centro de Investigación y de Estudios Avanzados del IPN · Tecnológico de Monterrey +2 more

Trust-based federated learning defense against Byzantine clients and data corruption in medical imaging classification tasks

Data Poisoning Attack visionfederated-learning
PDF
Loading more papers…