benchmark 2025

Black-Box Auditing of Quantum Model: Lifted Differential Privacy with Quantum Canaries

Baobao Song 1, Shiva Raj Pokhrel 2, Athanasios V. Vasilakos 3, Tianqing Zhu 4, Gang Li 2

0 citations · 26 references · arXiv

α

Published on arXiv

2512.14388

Membership Inference Attack

OWASP ML Top 10 — ML04

Key Finding

The proposed canary-based black-box framework successfully derives empirical lower bounds on privacy budget consumption in QML models, validated on real quantum hardware, bridging the gap between theoretical QDP guarantees and practical privacy verification.

Lifted Quantum Differential Privacy Auditing with Quantum Canaries

Novel technique introduced


Quantum machine learning (QML) promises significant computational advantages, yet models trained on sensitive data risk memorizing individual records, creating serious privacy vulnerabilities. While Quantum Differential Privacy (QDP) mechanisms provide theoretical worst-case guarantees, they critically lack empirical verification tools for deployed models. We introduce the first black-box privacy auditing framework for QML based on Lifted Quantum Differential Privacy, leveraging quantum canaries (strategically offset-encoded quantum states) to detect memorization and precisely quantify privacy leakage during training. Our framework establishes a rigorous mathematical connection between canary offset and trace distance bounds, deriving empirical lower bounds on privacy budget consumption that bridge the critical gap between theoretical guarantees and practical privacy verification. Comprehensive evaluations across both simulated and physical quantum hardware demonstrate our framework's effectiveness in measuring actual privacy loss in QML models, enabling robust privacy verification in QML systems.


Key Contributions

  • First black-box privacy auditing framework for QML using Lifted Quantum Differential Privacy and quantum canaries (offset-encoded quantum states) to detect memorization
  • Rigorous mathematical connection between canary offset and trace distance bounds, enabling empirical lower bounds on the privacy budget (ε) that bridge theory and practice
  • Comprehensive evaluation on both simulated and physical quantum hardware demonstrating the framework's effectiveness at measuring actual privacy loss

🛡️ Threat Analysis

Membership Inference Attack

Quantum canaries are strategically inserted training examples used to test whether the QML model memorized them — this is fundamentally a membership inference technique (binary: was this canary in the training set?), used here to derive empirical lower bounds on actual privacy budget consumption and audit DP guarantees in deployed QML models.


Details

Threat Tags
black_boxtraining_time
Applications
quantum machine learningprivacy auditing