tool 2026

ArtifactLens: Hundreds of Labels Are Enough for Artifact Detection with VLMs

James Burgess 1, Rameen Abdal 2, Dan Stoddart 2, Sergey Tulyakov 2, Serena Yeung-Levy 1, Kuan-Chieh Jackson Wang 2

0 citations · 85 references · arXiv (Cornell University)

α

Published on arXiv

2602.09475

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Achieves state-of-the-art on five human artifact benchmarks using orders of magnitude less labeled data (hundreds vs. tens of thousands of examples) compared to fine-tuned VLM baselines

ArtifactLens

Novel technique introduced


Modern image generators produce strikingly realistic images, where only artifacts like distorted hands or warped objects reveal their synthetic origin. Detecting these artifacts is essential: without detection, we cannot benchmark generators or train reward models to improve them. Current detectors fine-tune VLMs on tens of thousands of labeled images, but this is expensive to repeat whenever generators evolve or new artifact types emerge. We show that pretrained VLMs already encode the knowledge needed to detect artifacts - with the right scaffolding, this capability can be unlocked using only a few hundred labeled examples per artifact category. Our system, ArtifactLens, achieves state-of-the-art on five human artifact benchmarks (the first evaluation across multiple datasets) while requiring orders of magnitude less labeled data. The scaffolding consists of a multi-component architecture with in-context learning and text instruction optimization, with novel improvements to each. Our methods generalize to other artifact types - object morphology, animal anatomy, and entity interactions - and to the distinct task of AIGC detection.


Key Contributions

  • ArtifactLens: a multi-component scaffold around frozen VLMs using specialist sub-models, in-context learning, and black-box text instruction optimization, requiring only hundreds of labeled examples per artifact category
  • First evaluation of AI-generated image artifact detectors across five human artifact benchmarks simultaneously, achieving state-of-the-art results
  • Demonstrated generalization to object morphology, animal anatomy, entity interactions, and AIGC detection tasks

🛡️ Threat Analysis

Output Integrity Attack

The paper's primary contribution is an AI-generated content (AIGC) detection system — detecting synthetic image artifacts (distorted hands, warped objects) produced by generative models. This is directly about verifying the authenticity and origin of visual content, a core ML09 concern. The paper explicitly evaluates on the 'distinct task of AIGC detection' alongside artifact detection.


Details

Domains
visionmultimodalgenerative
Model Types
vlmdiffusiontransformer
Threat Tags
inference_timeblack_box
Datasets
five human artifact benchmarks (multi-dataset evaluation)
Applications
ai-generated image detectionimage artifact detectionimage quality assessment for generative models