attack 2026

Vulnerabilities in Partial TEE-Shielded LLM Inference with Precomputed Noise

Abhishek Saini , Haolin Jiang , Hang Liu

0 citations · 91 references · arXiv (Cornell University)

α

Published on arXiv

2602.11088

Model Theft

OWASP ML Top 10 — ML05

Model Theft

OWASP LLM Top 10 — LLM10

Key Finding

Recovers a LLaMA-3 8B model layer's secret basis in approximately 6 minutes, with the attack scaling to fully compromise 405B-parameter LLMs across a variety of configurations.

Cross-Query Key Reuse Attack

Novel technique introduced


The deployment of large language models (LLMs) on third-party devices requires new ways to protect model intellectual property. While Trusted Execution Environments (TEEs) offer a promising solution, their performance limits can lead to a critical compromise: using a precomputed, static secret basis to accelerate cryptographic operations. We demonstrate that this mainstream design pattern introduces a classic cryptographic flaw, the reuse of secret keying material, into the system's protocol. We prove its vulnerability with two distinct attacks: First, our attack on a model confidentiality system achieves a full confidentiality break by recovering its secret permutations and model weights. Second, our integrity attack completely bypasses the integrity checks of systems like Soter and TSQP. We demonstrate the practicality of our attacks against state-of-the-art LLMs, recovering a layer's secrets from a LLaMA-3 8B model in about 6 minutes and showing the attack scales to compromise 405B-parameter LLMs across a variety of configurations.


Key Contributions

  • Identifies that the mainstream design pattern of using a precomputed static secret basis in partial TEE-shielded LLM inference introduces a classic key-reuse cryptographic flaw
  • Confidentiality attack that fully recovers secret permutations and model weights by exploiting cross-query correlation of the static secret basis
  • Integrity attack that completely bypasses the integrity verification of deployed systems (Soter, TSQP), demonstrated to scale to 405B-parameter LLMs

🛡️ Threat Analysis

Model Theft

Both attacks target model intellectual property: the confidentiality attack recovers secret permutations and model weights from TEE-protected systems, and the integrity attack bypasses the verification checks meant to prevent tampering — the primary outcome is unauthorized extraction of model weights.


Details

Domains
nlp
Model Types
llmtransformer
Threat Tags
grey_boxinference_time
Applications
llm inference on third-party/edge devicestee-shielded confidential computing for llmson-premises enterprise llm deployment