Latest papers

4 papers
attack arXiv Mar 22, 2026 · 15d ago

Is Monitoring Enough? Strategic Agent Selection For Stealthy Attack in Multi-Agent Discussions

Qiuchi Xiang, Haoxuan Qu, Hossein Rahmani et al. · Lancaster University

Stealth attack on multi-agent LLM discussions that evades continuous anomaly monitoring through strategic agent selection and message crafting

Prompt Injection Excessive Agency nlpmultimodal
PDF
attack arXiv Jan 30, 2026 · 9w ago

Hide and Seek in Embedding Space: Geometry-based Steganography and Detection in Large Language Models

Charles Westphal, Keivan Navaie, Fernando E. Rosas · University College London · ML Alignment Theory Scholars +4 more

Maliciously LoRA-fine-tuned LLMs covertly exfiltrate prompt secrets via geometry-based steganography, detected via linear probes on internal activations

Model Poisoning Sensitive Information Disclosure nlp
PDF
defense arXiv Dec 15, 2025 · Dec 2025

Learning to Generate Cross-Task Unexploitable Examples

Haoxuan Qu, Qiuchi Xiang, Yujun Cai et al. · Lancaster University · The University of Queensland +2 more

Defends personal images from unauthorized ML training by generating cross-task imperceptible perturbations that make training data unlearnable across diverse vision tasks

Data Poisoning Attack vision
PDF
defense arXiv Nov 17, 2025 · Nov 2025

Privacy-Preserving Federated Learning from Partial Decryption Verifiable Threshold Multi-Client Functional Encryption

Minjie Wang, Jinguang Han, Weizhi Meng · Southeast University · Lancaster University

Verifiable threshold functional encryption for FL defends against gradient leakage and adversarial aggregator poisoning attacks

Model Inversion Attack Data Poisoning Attack federated-learning
PDF