attack 2025

SEGA: A Transferable Signed Ensemble Gaussian Black-Box Attack against No-Reference Image Quality Assessment Models

Yujia Liu 1, Dingquan Li 2, Zhixuan Li 3, Tiejun Huang 1

0 citations · 74 references · TPAMI

α

Published on arXiv

2509.18546

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

SEGA achieves superior transferability over existing methods in black-box attacks against NR-IQA models on the CLIVE dataset.

SEGA (Signed Ensemble Gaussian Attack)

Novel technique introduced


No-Reference Image Quality Assessment (NR-IQA) models play an important role in various real-world applications. Recently, adversarial attacks against NR-IQA models have attracted increasing attention, as they provide valuable insights for revealing model vulnerabilities and guiding robust system design. Some effective attacks have been proposed against NR-IQA models in white-box settings, where the attacker has full access to the target model. However, these attacks often suffer from poor transferability to unknown target models in more realistic black-box scenarios, where the target model is inaccessible. This work makes the first attempt to address the challenge of low transferability in attacking NR-IQA models by proposing a transferable Signed Ensemble Gaussian black-box Attack (SEGA). The main idea is to approximate the gradient of the target model by applying Gaussian smoothing to source models and ensembling their smoothed gradients. To ensure the imperceptibility of adversarial perturbations, SEGA further removes inappropriate perturbations using a specially designed perturbation filter mask. Experimental results on the CLIVE dataset demonstrate the superior transferability of SEGA, validating its effectiveness in enabling successful transfer-based black-box attacks against NR-IQA models.


Key Contributions

  • First transferable black-box adversarial attack targeting NR-IQA models (SEGA), approximating target model gradients via Gaussian smoothing over an ensemble of source models.
  • Signed ensemble gradient strategy that improves cross-model transferability beyond standard gradient averaging.
  • Perturbation filter mask that removes imperceptibility-violating components while maintaining attack effectiveness.

🛡️ Threat Analysis

Input Manipulation Attack

SEGA crafts imperceptible adversarial perturbations at inference time to manipulate NR-IQA model outputs. The core contribution is a novel gradient approximation technique (Gaussian smoothing + signed ensemble) for improving adversarial transferability in black-box settings — a classic input manipulation attack.


Details

Domains
vision
Model Types
cnntransformer
Threat Tags
black_boxinference_timetargeteddigital
Datasets
CLIVE
Applications
image quality assessment