SEGA: A Transferable Signed Ensemble Gaussian Black-Box Attack against No-Reference Image Quality Assessment Models
Yujia Liu 1, Dingquan Li 2, Zhixuan Li 3, Tiejun Huang 1
Published on arXiv
2509.18546
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
SEGA achieves superior transferability over existing methods in black-box attacks against NR-IQA models on the CLIVE dataset.
SEGA (Signed Ensemble Gaussian Attack)
Novel technique introduced
No-Reference Image Quality Assessment (NR-IQA) models play an important role in various real-world applications. Recently, adversarial attacks against NR-IQA models have attracted increasing attention, as they provide valuable insights for revealing model vulnerabilities and guiding robust system design. Some effective attacks have been proposed against NR-IQA models in white-box settings, where the attacker has full access to the target model. However, these attacks often suffer from poor transferability to unknown target models in more realistic black-box scenarios, where the target model is inaccessible. This work makes the first attempt to address the challenge of low transferability in attacking NR-IQA models by proposing a transferable Signed Ensemble Gaussian black-box Attack (SEGA). The main idea is to approximate the gradient of the target model by applying Gaussian smoothing to source models and ensembling their smoothed gradients. To ensure the imperceptibility of adversarial perturbations, SEGA further removes inappropriate perturbations using a specially designed perturbation filter mask. Experimental results on the CLIVE dataset demonstrate the superior transferability of SEGA, validating its effectiveness in enabling successful transfer-based black-box attacks against NR-IQA models.
Key Contributions
- First transferable black-box adversarial attack targeting NR-IQA models (SEGA), approximating target model gradients via Gaussian smoothing over an ensemble of source models.
- Signed ensemble gradient strategy that improves cross-model transferability beyond standard gradient averaging.
- Perturbation filter mask that removes imperceptibility-violating components while maintaining attack effectiveness.
🛡️ Threat Analysis
SEGA crafts imperceptible adversarial perturbations at inference time to manipulate NR-IQA model outputs. The core contribution is a novel gradient approximation technique (Gaussian smoothing + signed ensemble) for improving adversarial transferability in black-box settings — a classic input manipulation attack.