attack 2026

Generative Adversarial Perturbations with Cross-paradigm Transferability on Localized Crowd Counting

Alabi Mehzabin Anisha , Guangjing Wang , Sriram Chellappan

0 citations

α

Published on arXiv

2603.24821

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Achieves 7X increase in Mean Absolute Error compared to clean images while maintaining imperceptibility, successfully transferring across seven SOTA models

CrowdGen

Novel technique introduced


State-of-the-art crowd counting and localization are primarily modeled using two paradigms: density maps and point regression. Given the field's security ramifications, there is active interest in model robustness against adversarial attacks. Recent studies have demonstrated transferability across density-map-based approaches via adversarial patches, but cross-paradigm attacks (i.e., across both density map-based models and point regression-based models) remain unexplored. We introduce a novel adversarial framework that compromises both density map and point regression architectural paradigms through a comprehensive multi-task loss optimization. For point-regression models, we employ scene-density-specific high-confidence logit suppression; for density-map approaches, we use peak-targeted density map suppression. Both are combined with model-agnostic perceptual constraints to ensure that perturbations are effective and imperceptible to the human eye. Extensive experiments demonstrate the effectiveness of our attack, achieving on average a 7X increase in Mean Absolute Error compared to clean images while maintaining competitive visual quality, and successfully transferring across seven state-of-the-art crowd models with transfer ratios ranging from 0.55 to 1.69. Our approach strikes a balance between attack effectiveness and imperceptibility compared to state-of-the-art transferable attack strategies. The source code is available at https://github.com/simurgh7/CrowdGen


Key Contributions

  • First cross-paradigm adversarial attack framework effective against both density-map and point-regression crowd counting models
  • Multi-task loss optimization combining scene-density-specific logit suppression and peak-targeted density map suppression with perceptual constraints
  • Demonstrates transferability across seven state-of-the-art models with transfer ratios 0.55-1.69

🛡️ Threat Analysis

Input Manipulation Attack

Generates adversarial perturbations that cause misclassification/misprediction in crowd counting models at inference time. Uses gradient-based optimization with perceptual constraints to create imperceptible perturbations that transfer across multiple model architectures and paradigms.


Details

Domains
vision
Model Types
cnn
Threat Tags
inference_timedigitaluntargetedblack_box
Datasets
ShanghaiTechUCF-QNRFNWPU-Crowd
Applications
crowd countingcrowd localization