defense 2026

Anonymization-Enhanced Privacy Protection for Mobile GUI Agents: Available but Invisible

Lepeng Zhao , Zhenhua Zou , Shuo Li , Zhuotao Liu

1 citations · 30 references · arXiv (Cornell University)

α

Published on arXiv

2602.10139

Sensitive Information Disclosure

OWASP LLM Top 10 — LLM06

Key Finding

Framework substantially reduces privacy leakage across multiple MLLM models while incurring only modest task utility degradation, achieving the best observed privacy-utility trade-off among compared methods.

GUI Privacy Protection Framework (available-but-invisible anonymization)

Novel technique introduced


Mobile Graphical User Interface (GUI) agents have demonstrated strong capabilities in automating complex smartphone tasks by leveraging multimodal large language models (MLLMs) and system-level control interfaces. However, this paradigm introduces significant privacy risks, as agents typically capture and process entire screen contents, thereby exposing sensitive personal data such as phone numbers, addresses, messages, and financial information. Existing defenses either reduce UI exposure, obfuscate only task-irrelevant content, or rely on user authorization, but none can protect task-critical sensitive information while preserving seamless agent usability. We propose an anonymization-based privacy protection framework that enforces the principle of available-but-invisible access to sensitive data: sensitive information remains usable for task execution but is never directly visible to the cloud-based agent. Our system detects sensitive UI content using a PII-aware recognition model and replaces it with deterministic, type-preserving placeholders (e.g., PHONE_NUMBER#a1b2c) that retain semantic categories while removing identifying details. A layered architecture comprising a PII Detector, UI Transformer, Secure Interaction Proxy, and Privacy Gatekeeper ensures consistent anonymization across user instructions, XML hierarchies, and screenshots, mediates all agent actions over anonymized interfaces, and supports narrowly scoped local computations when reasoning over raw values is necessary. Extensive experiments on the AndroidLab and PrivScreen benchmarks show that our framework substantially reduces privacy leakage across multiple models while incurring only modest utility degradation, achieving the best observed privacy-utility trade-off among existing methods. Code available at: https://github.com/one-step-beh1nd/gui_privacy_protection


Key Contributions

  • PII-aware recognition model that detects sensitive UI content and replaces it with deterministic, type-preserving placeholders (e.g., PHONE_NUMBER#a1b2c) across screenshots, XML hierarchies, and user instructions
  • Layered architecture (PII Detector, UI Transformer, Secure Interaction Proxy, Privacy Gatekeeper) enforcing 'available-but-invisible' access so agents can act on anonymized data without ever seeing raw PII
  • Empirical evaluation on AndroidLab and PrivScreen showing best privacy-utility trade-off among existing methods across multiple MLLM backends

🛡️ Threat Analysis


Details

Domains
multimodalnlp
Model Types
llmvlmmultimodal
Threat Tags
inference_timeblack_box
Datasets
AndroidLabPrivScreen
Applications
mobile gui agentssmartphone task automationmllm-based ui agents