Latest papers

2 papers
attack arXiv Nov 21, 2025 · Nov 2025

MURMUR: Using cross-user chatter to break collaborative language agents in groups

Atharv Singh Patlan, Peiyao Sheng, S. Ashwin Hebbar et al. · Princeton University · Sentient

Discovers cross-user poisoning: adversarial messages in shared LLM agent history hijack actions of other users at inference time

Prompt Injection Excessive Agency nlp
PDF
attack arXiv Sep 30, 2025 · Sep 2025

Are Robust LLM Fingerprints Adversarially Robust?

Anshul Nasery, Edoardo Contente, Alkin Kaz et al. · University of Washington · Sentient +1 more

Adaptive attacks bypass ten LLM fingerprinting schemes with near-perfect success by exploiting four systemic vulnerabilities in ownership verification

Model Theft Model Theft nlp
3 citations PDF