tool arXiv Oct 27, 2025 · Oct 2025
Luca Melis, Matthew Grange, Iden Kalemaj et al. · Meta
Open-source PyTorch tool for auditing ML model privacy via membership inference, reconstruction, and extraction attacks
Membership Inference Attack Model Inversion Attack visionnlpgenerative
The increasing deployment of Machine Learning (ML) models in sensitive domains motivates the need for robust, practical privacy assessment tools. PrivacyGuard is a comprehensive tool for empirical differential privacy (DP) analysis, designed to evaluate privacy risks in ML models through state-of-the-art inference attacks and advanced privacy measurement techniques. To this end, PrivacyGuard implements a diverse suite of privacy attack -- including membership inference , extraction, and reconstruction attacks -- enabling both off-the-shelf and highly configurable privacy analyses. Its modular architecture allows for the seamless integration of new attacks, and privacy metrics, supporting rapid adaptation to emerging research advances. We make PrivacyGuard available at https://github.com/facebookresearch/PrivacyGuard.
llm transformer traditional_ml Meta
benchmark arXiv Nov 18, 2025 · Nov 2025
Iden Kalemaj, Luca Melis, Maxime Boucher et al. · Meta
Observational DP auditing framework verifies label and attribute privacy without modifying training data, extending beyond membership inference
Model Inversion Attack Membership Inference Attack visiontabular
Differential privacy (DP) auditing is essential for evaluating privacy guarantees in machine learning systems. Existing auditing methods, however, pose a significant challenge for large-scale systems since they require modifying the training dataset -- for instance, by injecting out-of-distribution canaries or removing samples from training. Such interventions on the training data pipeline are resource-intensive and involve considerable engineering overhead. We introduce a novel observational auditing framework that leverages the inherent randomness of data distributions, enabling privacy evaluation without altering the original dataset. Our approach extends privacy auditing beyond traditional membership inference to protected attributes, with labels as a special case, addressing a key gap in existing techniques. We provide theoretical foundations for our method and perform experiments on Criteo and CIFAR-10 datasets that demonstrate its effectiveness in auditing label privacy guarantees. This work opens new avenues for practical privacy auditing in large-scale production environments.
cnn traditional_ml Meta