defense 2026

Serverless AI Security: Attack Surface Analysis and Runtime Protection Mechanisms for FaaS-Based Machine Learning

Chetan Pathade , Vinod Dhimam , Sheheryar Ahmad , Ilsa Lareb

0 citations · 61 references · arXiv

α

Published on arXiv

2601.11664

AI Supply Chain Attacks

OWASP ML Top 10 — ML06

Model Theft

OWASP ML Top 10 — ML05

Key Finding

SAS achieves 94% threat detection rate while maintaining inference latency overhead below 9% across three major FaaS platforms

Serverless AI Shield (SAS)

Novel technique introduced


Serverless computing has achieved widespread adoption, with over 70% of AWS organizations using serverless solutions [1]. Meanwhile, machine learning inference workloads increasingly migrate to Function-as-a-Service (FaaS) platforms for their scalability and cost-efficiency [2], [3], [4]. However, this convergence introduces critical security challenges, with recent reports showing a 220% increase in AI/ML vulnerabilities [5] and serverless computing's fragmented architecture raises new security concerns distinct from traditional cloud deployments [6], [7]. This paper presents the first comprehensive security analysis of machine learning workloads in serverless environments. We systematically characterize the attack surface across five categories: function-level vulnerabilities (cold start exploitation, dependency poisoning), model-specific threats (API-based extraction, adversarial inputs), infrastructure attacks (cross-function contamination, privilege escalation), supply chain risks (malicious layers, backdoored libraries), and IAM complexity (ephemeral nature, serverless functions). Through empirical assessments across AWS Lambda, Azure Functions, and Google Cloud Functions, we demonstrate real-world attack scenarios and quantify their security impact. We propose Serverless AI Shield (SAS), a multi-layered defense framework providing pre-deployment validation, runtime monitoring, and post-execution forensics. Our evaluation shows SAS achieves 94% detection rates while maintaining performance overhead below 9% for inference latency. We release an open-source security toolkit to enable practitioners to assess and harden their serverless AI deployments, advancing the field toward more resilient cloud-native machine learning systems.


Key Contributions

  • First comprehensive attack surface characterization for ML workloads in serverless/FaaS environments across five categories: function-level vulnerabilities, model-specific threats, infrastructure attacks, supply chain risks, and IAM complexity
  • Serverless AI Shield (SAS): a multi-layered defense framework combining pre-deployment validation, runtime monitoring, and post-execution forensics achieving 94% detection rate with under 9% inference latency overhead
  • Open-source security toolkit empirically validated across AWS Lambda, Azure Functions, and Google Cloud Functions for assessing and hardening serverless AI deployments

🛡️ Threat Analysis

Model Theft

API-based model extraction is enumerated as a key model-specific threat in serverless ML deployments, where ephemeral function endpoints expose model inference APIs to adversarial querying for cloning.

AI Supply Chain Attacks

Explicitly analyzes supply chain attack vectors in serverless ML pipelines — malicious Lambda layers, backdoored libraries, and dependency poisoning — as a primary threat category unique to FaaS-based ML deployments.


Details

Threat Tags
black_boxinference_timetraining_time
Applications
machine learning inferenceserverless ml pipelinesfaas-based ai workloads