Hidden Tail: Adversarial Image Causing Stealthy Resource Consumption in Vision-Language Models
Rui Zhang 1, Zihan Wang 1, Tianli Yang 1, Hongwei Li 1, Wenbo Jiang 1, Qingchuan Zhao 2, Yang Liu 3, Guowen Xu 1
Published on arXiv
2508.18805
Input Manipulation Attack
OWASP ML Top 10 — ML01
Model Denial of Service
OWASP LLM Top 10 — LLM04
Key Finding
Hidden Tail increases VLM output length by up to 19.2x, reaching the maximum token limit, while preserving attack stealthiness with user-visible outputs indistinguishable from clean image responses.
Hidden Tail
Novel technique introduced
Vision-Language Models (VLMs) are increasingly deployed in real-world applications, but their high inference cost makes them vulnerable to resource consumption attacks. Prior attacks attempt to extend VLM output sequences by optimizing adversarial images, thereby increasing inference costs. However, these extended outputs often introduce irrelevant abnormal content, compromising attack stealthiness. This trade-off between effectiveness and stealthiness poses a major limitation for existing attacks. To address this challenge, we propose \textit{Hidden Tail}, a stealthy resource consumption attack that crafts prompt-agnostic adversarial images, inducing VLMs to generate maximum-length outputs by appending special tokens invisible to users. Our method employs a composite loss function that balances semantic preservation, repetitive special token induction, and suppression of the end-of-sequence (EOS) token, optimized via a dynamic weighting strategy. Extensive experiments show that \textit{Hidden Tail} outperforms existing attacks, increasing output length by up to 19.2$\times$ and reaching the maximum token limit, while preserving attack stealthiness. These results highlight the urgent need to improve the robustness of VLMs against efficiency-oriented adversarial threats. Our code is available at https://github.com/zhangrui4041/Hidden_Tail.
Key Contributions
- Hidden Tail attack: prompt-agnostic adversarial images that induce VLMs to generate maximum-length outputs via user-invisible special tokens appended as a hidden tail
- Composite loss function balancing semantic consistency, hidden tail induction, and EOS suppression with dynamic weighting strategy
- Demonstrates 19.2x output length increase reaching maximum token limit while maintaining user-visible output quality comparable to clean images
🛡️ Threat Analysis
Crafts adversarial images via gradient-based optimization (composite loss with EOS suppression, hidden tail induction) that manipulate VLM inference-time behavior — this is a visual adversarial input attack on a multimodal model.