attack 2026

Membership Inference for Contrastive Pre-training Models with Text-only PII Queries

Ruoxi Cheng 1,2, Yizhong Ding 1, Hongyi Zhang 3, Yiyan Huang 4

0 citations

α

Published on arXiv

2603.14222

Membership Inference Attack

OWASP ML Top 10 — ML04

Key Finding

Achieves strong membership inference detection performance with sub-second auditing cost while using only text queries, avoiding direct exposure of sensitive biometric information

UMID

Novel technique introduced


Contrastive pretraining models such as CLIP and CLAP underpin many vision-language and audio-language systems, yet their reliance on web-scale data raises growing concerns about memorizing Personally Identifiable Information (PII). Auditing such models via membership inference is challenging in practice: shadow-model MIAs are computationally prohibitive for large multimodal backbones, and existing multimodal attacks typically require querying the target with paired biometric inputs, thereby directly exposing sensitive biometric information to the target model. We propose Unimodal Membership Inference Detector (UMID), a text-only auditing framework that performs text-guided cross-modal latent inversion and extracts two complementary signals, similarity (alignment to the queried text) and variability (consistency across randomized inversions). UMID compares these statistics to a lightweight non-member reference constructed from synthetic gibberish and makes decisions via an ensemble of unsupervised anomaly detectors. Comprehensive experiments across diverse CLIP and CLAP architectures demonstrate that UMID significantly improves the effectiveness and efficiency over prior MIAs, delivering strong detection performance with sub-second auditing cost while complying with realistic privacy constraints.


Key Contributions

  • UMID framework performing text-only membership inference on multimodal contrastive models without exposing biometric data
  • Text-guided cross-modal latent inversion extracting similarity and variability signals for membership detection
  • Ensemble of unsupervised anomaly detectors using synthetic gibberish as non-member reference, achieving sub-second auditing cost

🛡️ Threat Analysis

Membership Inference Attack

Core contribution is a membership inference attack determining whether specific PII data points were in the training set of CLIP/CLAP models. The attack queries models with text descriptions of PII (e.g., person names) and uses cross-modal inversion with anomaly detection to infer membership, achieving this without requiring paired biometric inputs.


Details

Domains
multimodalvisionaudionlp
Model Types
multimodaltransformer
Threat Tags
black_boxinference_time
Datasets
CLIPCLAP
Applications
vision-language modelsaudio-language modelspii auditing