Latest papers

3 papers
attack Artificial Intelligence and fO... Oct 1, 2025 · Oct 2025

Attack logics, not outputs: Towards efficient robustification of deep neural networks by falsifying concept-based properties

Raik Dankworth, Gesina Schwalbe · University of Lübeck

Proposes adversarial attacks that falsify concept-based logical properties (e.g., red∧octagonal→stop_sign) rather than just output class labels, for more efficient NN robustification

Input Manipulation Attack vision
PDF
attack arXiv Sep 11, 2025 · Sep 2025

Prompt Pirates Need a Map: Stealing Seeds helps Stealing Prompts

Felix Mächtle, Ashwath Shetty, Jonas Sander et al. · University of Lübeck · Kiel University

Exploits PyTorch's 32-bit seed space to brute-force generation seeds and steal prompts from diffusion model outputs

Model Inversion Attack generativevision
PDF
attack arXiv Aug 7, 2025 · Aug 2025

Non-omniscient backdoor injection with one poison sample: Proving the one-poison hypothesis for linear regression, linear classification, and 2-layer ReLU neural networks

Thorsten Peinemann, Paula Arnold, Sebastian Berndt et al. · University of Lübeck · Technische Hochschule Lübeck

Proves one poison sample suffices to backdoor linear models and 2-layer ReLU networks with zero backdoor error and no full data knowledge

Model Poisoning
PDF