Leo Schwinn

h-index: 12 661 citations 54 papers (total)

Papers in Database (2)

attack arXiv Oct 31, 2025 · Oct 2025

Diffusion LLMs are Natural Adversaries for any LLM

David Lüdke, Tom Wollschläger, Paul Ungermann et al. · Technical University of Munich

Uses Diffusion LLMs as amortized jailbreak generators, producing low-perplexity transferable harmful prompts against black-box and proprietary LLMs

Prompt Injection nlpgenerative
3 citations PDF Code
tool arXiv Nov 6, 2025 · Nov 2025

AdversariaLLM: A Unified and Modular Toolbox for LLM Robustness Research

Tim Beyer, Jonas Dornbusch, Jakob Steimle et al. · Technical University of Munich · Munich Data Science Institute

Unified toolbox for reproducible LLM jailbreak research implementing 12 attacks, 7 datasets, and 13 judges

Input Manipulation Attack Prompt Injection nlp
2 citations PDF Code