Latest papers

4 papers
benchmark arXiv Feb 11, 2026 · 7w ago

Generative clinical time series models trained on moderate amounts of patient data are privacy preserving

Rustam Zhumagambetov, Niklas Giesa, Sebastian D. Boie et al. · Physikalisch-Technische Bundesanstalt (PTB) · Charité – Universitätsmedizin Berlin +1 more

Audits generative clinical time series models with membership inference and reconstruction attacks, finding large training sets confer natural privacy protection

Membership Inference Attack Model Inversion Attack timeseries
PDF
benchmark arXiv Feb 10, 2026 · 7w ago

Stop Testing Attacks, Start Diagnosing Defenses: The Four-Checkpoint Framework Reveals Where LLM Safety Breaks

Hayfa Dhahbi, Kashyap Thimmaraju · Technische Universität Berlin

Proposes Four-Checkpoint Framework and WASR metric to diagnose which LLM safety layers break under 13 prompt-level jailbreak techniques

Prompt Injection nlp
PDF
defense arXiv Aug 28, 2025 · Aug 2025

Towards Mechanistic Defenses Against Typographic Attacks in CLIP

Lorenz Hufe, Constantin Venhoff, Erblina Purelku et al. · Fraunhofer Heinrich Hertz Institute · University of Oxford +2 more

Defends CLIP against typographic image-text attacks via gradient-free attention head ablation, improving robustness 22% with <1% accuracy loss

Input Manipulation Attack Prompt Injection visionmultimodal
PDF Code
defense arXiv Aug 18, 2025 · Aug 2025

Beyond Trade-offs: A Unified Framework for Privacy, Robustness, and Communication Efficiency in Federated Learning

Yue Xia, Tayyebeh Jahani-Nezhad, Rawad Bitar · Technical University of Munich · Technische Universität Berlin

Defends federated learning against Byzantine clients using JL-compression-compatible robust aggregation with differential privacy guarantees

Data Poisoning Attack federated-learning
PDF