benchmark 2025

DiffSeg30k: A Multi-Turn Diffusion Editing Benchmark for Localized AIGC Detection

Hai Ci 1, Ziheng Peng 2, Pei Yang 1, Yingxin Xuan 1, Mike Zheng Shou 1

0 citations · 73 references · arXiv

α

Published on arXiv

2511.19111

Output Integrity Attack

OWASP ML Top 10 — ML09

Key Finding

Segmentation models trained for pixel-level localization emerge as more reliable whole-image classifiers of diffusion edits than established forgery classifiers, despite not being explicitly trained for binary classification.

DiffSeg30k

Novel technique introduced


Diffusion-based editing enables realistic modification of local image regions, making AI-generated content harder to detect. Existing AIGC detection benchmarks focus on classifying entire images, overlooking the localization of diffusion-based edits. We introduce DiffSeg30k, a publicly available dataset of 30k diffusion-edited images with pixel-level annotations, designed to support fine-grained detection. DiffSeg30k features: 1) In-the-wild images--we collect images or image prompts from COCO to reflect real-world content diversity; 2) Diverse diffusion models--local edits using eight SOTA diffusion models; 3) Multi-turn editing--each image undergoes up to three sequential edits to mimic real-world sequential editing; and 4) Realistic editing scenarios--a vision-language model (VLM)-based pipeline automatically identifies meaningful regions and generates context-aware prompts covering additions, removals, and attribute changes. DiffSeg30k shifts AIGC detection from binary classification to semantic segmentation, enabling simultaneous localization of edits and identification of the editing models. We benchmark three baseline segmentation approaches, revealing significant challenges in semantic segmentation tasks, particularly concerning robustness to image distortions. Experiments also reveal that segmentation models, despite being trained for pixel-level localization, emerge as highly reliable whole-image classifiers of diffusion edits, outperforming established forgery classifiers while showing great potential in cross-generator generalization. We believe DiffSeg30k will advance research in fine-grained localization of AI-generated content by demonstrating the promise and limitations of segmentation-based methods. DiffSeg30k is released at: https://huggingface.co/datasets/Chaos2629/Diffseg30k


Key Contributions

  • DiffSeg30k: a publicly released dataset of 30k diffusion-edited images with pixel-level edit annotations, sourced from COCO using 8 SOTA diffusion models and up to three sequential edits per image
  • VLM-based pipeline that automatically identifies meaningful image regions and generates context-aware editing prompts covering additions, removals, and attribute changes for realistic edit simulation
  • Benchmarking of segmentation-based AIGC detection, revealing that segmentation models outperform established binary forgery classifiers and showing strong cross-generator generalization potential

🛡️ Threat Analysis

Output Integrity Attack

Core contribution is enabling fine-grained detection of AI-generated/diffusion-edited content — shifting AIGC detection from binary image classification to pixel-level semantic segmentation for output integrity verification and content provenance.


Details

Domains
visiongenerative
Model Types
diffusionvlm
Threat Tags
inference_time
Datasets
DiffSeg30kCOCO
Applications
ai-generated content detectionimage forensicsdiffusion edit localization