defense 2025

Defending Diffusion Models Against Membership Inference Attacks via Higher-Order Langevin Dynamics

Benjamin Sterling 1, Yousef El-Laham 2, Mónica F. Bugallo 2

0 citations

α

Published on arXiv

2509.14225

Membership Inference Attack

OWASP ML Top 10 — ML04

Key Finding

HOLD++ theoretically and empirically reduces MIA success (measured by AUROC) on diffusion models trained on speech data while maintaining sample quality as measured by FID

HOLD++ (Critically-Damped Higher-Order Langevin Dynamics)

Novel technique introduced


Recent advances in generative artificial intelligence applications have raised new data security concerns. This paper focuses on defending diffusion models against membership inference attacks. This type of attack occurs when the attacker can determine if a certain data point was used to train the model. Although diffusion models are intrinsically more resistant to membership inference attacks than other generative models, they are still susceptible. The defense proposed here utilizes critically-damped higher-order Langevin dynamics, which introduces several auxiliary variables and a joint diffusion process along these variables. The idea is that the presence of auxiliary variables mixes external randomness that helps to corrupt sensitive input data earlier on in the diffusion process. This concept is theoretically investigated and validated on a toy dataset and a speech dataset using the Area Under the Receiver Operating Characteristic (AUROC) curves and the FID metric.


Key Contributions

  • Applies critically-damped higher-order Langevin dynamics (HOLD++) to diffusion model training as a defense against membership inference attacks by introducing auxiliary variables that mix external randomness and corrupt sensitive inputs earlier in the forward diffusion process
  • Theoretical analysis showing HOLD++ achieves a base level of differential privacy and defends against score-network-based MIA methods such as SecMI and PIA
  • Empirical validation on a toy dataset and a speech dataset using AUROC curves and FID, demonstrating the privacy-utility tradeoff of the proposed approach

🛡️ Threat Analysis

Membership Inference Attack

The paper's explicit and sole security goal is defending against membership inference attacks on diffusion models — the attacker's goal is to determine whether a specific data point was used in training, which is the canonical ML04 threat.


Details

Domains
generativeaudio
Model Types
diffusion
Threat Tags
black_boxtraining_time
Datasets
toy datasetspeech dataset
Applications
generative audio synthesisdiffusion model training with sensitive datamedical data generation