PRIVMARK: Private Large Language Models Watermarking with MPC
Thomas Fargues 1, Ye Dong 2, Tianwei Zhang 3, Jin-Song Dong 2
Published on arXiv
2509.24624
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
PRIVMARK achieves semantically identical watermarking results compared to plaintext PostMark while preserving model weight privacy across multiple computing parties and resisting paraphrasing and removal attacks
PRIVMARK
Novel technique introduced
The rapid growth of Large Language Models (LLMs) has highlighted the pressing need for reliable mechanisms to verify content ownership and ensure traceability. Watermarking offers a promising path forward, but it remains limited by privacy concerns in sensitive scenarios, as traditional approaches often require direct access to a model's parameters or its training data. In this work, we propose a secure multi-party computation (MPC)-based private LLMs watermarking framework, PRIVMARK, to address the concerns. Concretely, we investigate PostMark (EMNLP'2024), one of the state-of-the-art LLMs Watermarking methods, and formulate its basic operations. Then, we construct efficient protocols for these operations using the MPC primitives in a black-box manner. In this way, PRIVMARK enables multiple parties to collaboratively watermark an LLM's output without exposing the model's weights to any single computing party. We implement PRIVMARK using SecretFlow-SPU (USENIX ATC'2023) and evaluate its performance using the ABY3 (CCS'2018) backend. The experimental results show that PRIVMARK achieves semantically identical results compared to the plaintext baseline without MPC and is resistant against paraphrasing and removing attacks with reasonable efficiency.
Key Contributions
- Formalizes PostMark's operations and constructs efficient MPC-compatible protocols for each primitive in a black-box manner using SecretFlow-SPU and ABY3
- Enables multiple parties to collaboratively watermark LLM outputs without any single party seeing model weights, addressing privacy in multi-party LLM deployments
- Demonstrates that PRIVMARK achieves semantically identical watermarking quality to the plaintext baseline while resisting paraphrasing and watermark removal attacks
🛡️ Threat Analysis
Embeds watermarks in LLM-generated text outputs for content provenance and traceability; the watermark is in the text OUTPUT (not model weights), and the paper explicitly evaluates resistance to watermark removal and paraphrasing attacks — squarely output integrity and content authentication.