survey arXiv Nov 13, 2025 · Nov 2025
Josep Domingo-Ferrer · Universitat Rovira i Virgili · LAAS-CNRS
Surveys real-world effectiveness of MIA, property inference, and reconstruction attacks, arguing current regulatory concerns are overstated
Membership Inference Attack Model Inversion Attack tabulargenerativevision
In several jurisdictions, the regulatory framework on the release and sharing of personal data is being extended to machine learning (ML). The implicit assumption is that disclosing a trained ML model entails a privacy risk for any personal data used in training comparable to directly releasing those data. However, given a trained model, it is necessary to mount a privacy attack to make inferences on the training data. In this concept paper, we examine the main families of privacy attacks against predictive and generative ML, including membership inference attacks (MIAs), property inference attacks, and reconstruction attacks. Our discussion shows that most of these attacks seem less effective in the real world than what a prima face interpretation of the related literature could suggest.
traditional_ml gan diffusion transformer Universitat Rovira i Virgili · LAAS-CNRS
defense arXiv Nov 20, 2025 · Nov 2025
Mona Khalil, Alberto Blanco-Justicia, Najeeb Jebreel et al. · Universitat Rovira i Virgili · LAAS-CNRS
Analyzes why outlier training samples remain vulnerable to membership inference attacks even in non-overfitted models and proposes defenses
Membership Inference Attack vision
Membership inference attacks (MIAs) against machine learning (ML) models aim to determine whether a given data point was part of the model training data. These attacks may pose significant privacy risks to individuals whose sensitive data were used for training, which motivates the use of defenses such as differential privacy, often at the cost of high accuracy losses. MIAs exploit the differences in the behavior of a model when making predictions on samples it has seen during training (members) versus those it has not seen (non-members). Several studies have pointed out that model overfitting is the major factor contributing to these differences in behavior and, consequently, to the success of MIAs. However, the literature also shows that even non-overfitted ML models can leak information about a small subset of their training data. In this paper, we investigate the root causes of membership inference vulnerabilities beyond traditional overfitting concerns and suggest targeted defenses. We empirically analyze the characteristics of the training data samples vulnerable to MIAs in models that are not overfitted (and hence able to generalize). Our findings reveal that these samples are often outliers within their classes (e.g., noisy or hard to classify). We then propose potential defensive strategies to protect these vulnerable samples and enhance the privacy-preserving capabilities of ML models. Our code is available at https://github.com/najeebjebreel/mia_analysis.
cnn transformer Universitat Rovira i Virgili · LAAS-CNRS