defense 2025

Secure Confidential Business Information When Sharing Machine Learning Models

Yunfan Yang 1, Jiarong Xu 1, Hongzhe Zhang 2, Xiao Fang 3

0 citations

α

Published on arXiv

2509.16352

Model Inversion Attack

OWASP ML Top 10 — ML03

Key Finding

The proposed defense outperforms existing CPI defenses in resisting adaptive property inference attacks while better preserving model utility and reducing computational overhead across realistic model-sharing scenarios

Responsive CPI Arms Race Defense

Novel technique introduced


Model-sharing offers significant business value by enabling firms with well-established Machine Learning (ML) models to monetize and share their models with others who lack the resources to develop ML models from scratch. However, concerns over data confidentiality remain a significant barrier to model-sharing adoption, as Confidential Property Inference (CPI) attacks can exploit shared ML models to uncover confidential properties of the model provider's private model training data. Existing defenses often assume that CPI attacks are non-adaptive to the specific ML model they are targeting. This assumption overlooks a key characteristic of real-world adversaries: their responsiveness, i.e., adversaries' ability to dynamically adjust their attack models based on the information of the target and its defenses. To overcome this limitation, we propose a novel defense method that explicitly accounts for the responsive nature of real-world adversaries via two methodological innovations: a novel Responsive CPI attack and an attack-defense arms race framework. The former emulates the responsive behaviors of adversaries in the real world, and the latter iteratively enhances both the target and attack models, ultimately producing a secure ML model that is robust against responsive CPI attacks. Furthermore, we propose and integrate a novel approximate strategy into our defense, which addresses a critical computational bottleneck of defense methods and improves defense efficiency. Through extensive empirical evaluations across various realistic model-sharing scenarios, we demonstrate that our method outperforms existing defenses by more effectively defending against CPI attacks, preserving ML model utility, and reducing computational overhead.


Key Contributions

  • Novel 'Responsive CPI attack' that emulates adaptive adversaries who dynamically adjust their attack strategy based on information about the target model and its deployed defenses
  • Attack-defense arms race framework that iteratively co-evolves both the target model and the attack model, producing a shared ML model robust against responsive property inference attacks
  • Approximate strategy that addresses the computational bottleneck in the arms race defense loop, significantly reducing overhead while preserving defense quality

🛡️ Threat Analysis

Model Inversion Attack

CPI attacks recover confidential private attributes of the model provider's training data from a shared ML model — the adversary infers sensitive training data properties by querying the model, which squarely falls under recovering private attributes from a trained model. The defense is explicitly evaluated against this reconstruction/inference threat.


Details

Model Types
traditional_ml
Threat Tags
black_boxgrey_boxinference_time
Applications
model sharingml model monetizationbusiness ml apis