attack 2025

DCMI: A Differential Calibration Membership Inference Attack Against Retrieval-Augmented Generation

Xinyu Gao , Xiangtao Meng , Yingkai Dong , Zheng Li , Shanqing Guo

0 citations

α

Published on arXiv

2509.06026

Membership Inference Attack

OWASP ML Top 10 — ML04

Sensitive Information Disclosure

OWASP LLM Top 10 — LLM06

Key Finding

DCMI achieves 97.42% AUC and 94.35% accuracy against Flan-T5-based RAG, exceeding the MBA baseline by over 40%, with 10–20% gains on real-world RAG platforms

DCMI

Novel technique introduced


While Retrieval-Augmented Generation (RAG) effectively reduces hallucinations by integrating external knowledge bases, it introduces vulnerabilities to membership inference attacks (MIAs), particularly in systems handling sensitive data. Existing MIAs targeting RAG's external databases often rely on model responses but ignore the interference of non-member-retrieved documents on RAG outputs, limiting their effectiveness. To address this, we propose DCMI, a differential calibration MIA that mitigates the negative impact of non-member-retrieved documents. Specifically, DCMI leverages the sensitivity gap between member and non-member retrieved documents under query perturbation. It generates perturbed queries for calibration to isolate the contribution of member-retrieved documents while minimizing the interference from non-member-retrieved documents. Experiments under progressively relaxed assumptions show that DCMI consistently outperforms baselines--for example, achieving 97.42% AUC and 94.35% Accuracy against the RAG system with Flan-T5, exceeding the MBA baseline by over 40%. Furthermore, on real-world RAG platforms such as Dify and MaxKB, DCMI maintains a 10%-20% advantage over the baseline. These results highlight significant privacy risks in RAG systems and emphasize the need for stronger protection mechanisms. We appeal to the community's consideration of deeper investigations, like ours, against the data leakage risks in rapidly evolving RAG systems. Our code is available at https://github.com/Xinyu140203/RAG_MIA.


Key Contributions

  • Identifies that non-member retrieved documents interfere with and degrade existing MIAs targeting RAG knowledge bases
  • Proposes DCMI which exploits the sensitivity gap between member and non-member documents under query perturbation to isolate member-document contribution signals
  • Demonstrates 97.42% AUC on Flan-T5 RAG systems (40%+ over MBA baseline) and 10–20% advantage on real-world platforms Dify and MaxKB

🛡️ Threat Analysis

Membership Inference Attack

Core contribution is a membership inference attack that determines whether a specific document is in a RAG system's retrieval database — the canonical ML04 threat, applied to RAG external knowledge stores rather than model training sets.


Details

Domains
nlp
Model Types
llmtransformer
Threat Tags
black_boxinference_time
Datasets
DifyMaxKB
Applications
retrieval-augmented generationrag knowledge baseshealthcare data systemsfinancial data systemslegal document systems