Privacy-Preserving Mechanisms Enable Cheap Verifiable Inference of LLMs
Arka Pal 1,2, Louai Zahran 1,2, William Gvozdjak 1,2,3, Akilesh Potti 1,2, Micah Goldblum 1,2,3
Published on arXiv
2602.17223
Output Integrity Attack
OWASP ML Top 10 — ML09
Key Finding
Privacy-preserving verifiable inference protocol using SMPC (SIGMA) runs ~15x faster than state-of-the-art ZK methods for Llama-2-7B with ~99% transcription success and negligible downstream impact.
Logit Fingerprinting with Noise
Novel technique introduced
As large language models (LLMs) continue to grow in size, fewer users are able to host and run models locally. This has led to increased use of third-party hosting services. However, in this setting, there is a lack of guarantees on the computation performed by the inference provider. For example, a dishonest provider may replace an expensive large model with a cheaper-to-run weaker model and return the results from the weaker model to the user. Existing tools to verify inference typically rely on methods from cryptography such as zero-knowledge proofs (ZKPs), but these add significant computational overhead, and remain infeasible for use for large models. In this work, we develop a new insight -- that given a method for performing private LLM inference, one can obtain forms of verified inference at marginal extra cost. Specifically, we propose two new protocols which leverage privacy-preserving LLM inference in order to provide guarantees over the inference that was carried out. Our approaches are cheap, requiring the addition of a few extra tokens of computation, and have little to no downstream impact. As the fastest privacy-preserving inference methods are typically faster than ZK methods, the proposed protocols also improve verification runtime. Our work provides novel insights into the connections between privacy and verifiability in LLM inference.
Key Contributions
- Novel insight that privacy-preserving inference mechanisms (SMPC, FHE) can be leveraged to provide verifiable inference guarantees at marginal extra cost
- Two concrete protocols: secret-key transcription and logit fingerprinting with noise, offering different levels of verification guarantees
- ~15x speedup over state-of-the-art ZK verification methods on Llama-2-7B with ~99% transcription rates and ~10^-15 guessing probability for secret keys
🛡️ Threat Analysis
The paper's primary contribution is verifiable inference schemes — protocols that prove model outputs actually came from the claimed LLM and were not substituted by a cheaper model. This directly targets output integrity and authenticity of LLM inference, which ML09 explicitly covers under 'verifiable inference schemes (proving outputs weren't tampered with)'.