attack 2026

Red-teaming the Multimodal Reasoning: Jailbreaking Vision-Language Models via Cross-modal Entanglement Attacks

Yu Yan 1,2, Sheng Sun 1, Shengjia Cheng 3, Teli Liu 3, Mingfeng Li 3, Min Liu 1,2

0 citations · arXiv (Cornell University)

α

Published on arXiv

2602.10148

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

COMET achieves state-of-the-art attack success rate against advanced VLMs' safety alignment under black-box conditions

COMET (CrossTALK)

Novel technique introduced


Vision-Language Models (VLMs) with multimodal reasoning capabilities are high-value attack targets, given their potential for handling complex multimodal harmful tasks. Mainstream black-box jailbreak attacks on VLMs work by distributing malicious clues across modalities to disperse model attention and bypass safety alignment mechanisms. However, these adversarial attacks rely on simple and fixed image-text combinations that lack attack complexity scalability, limiting their effectiveness for red-teaming VLMs' continuously evolving reasoning capabilities. We propose \textbf{CrossTALK} (\textbf{\underline{Cross}}-modal en\textbf{\underline{TA}}ng\textbf{\underline{L}}ement attac\textbf{\underline{K}}), which is a scalable approach that extends and entangles information clues across modalities to exceed VLMs' trained and generalized safety alignment patterns for jailbreak. Specifically, {knowledge-scalable reframing} extends harmful tasks into multi-hop chain instructions, {cross-modal clue entangling} migrates visualizable entities into images to build multimodal reasoning links, and {cross-modal scenario nesting} uses multimodal contextual instructions to steer VLMs toward detailed harmful outputs. Experiments show our COMET achieves state-of-the-art attack success rate.


Key Contributions

  • CrossTALK / COMET: a scalable cross-modal jailbreak framework that entangles harmful intent across text and image modalities to exceed VLM safety alignment patterns
  • Knowledge-scalable reframing that expands harmful tasks into multi-hop chain-of-thought instructions to increase attack complexity
  • Cross-modal clue entangling and cross-modal scenario nesting that migrate visualizable entities into images and use multimodal contextual instructions to steer VLMs toward detailed harmful outputs

🛡️ Threat Analysis


Details

Domains
multimodalvisionnlp
Model Types
vlmmultimodal
Threat Tags
black_boxinference_timetargeted
Applications
vision-language modelsmultimodal reasoning systems