attack 2025

Learning Fourier shapes to probe the geometric world of deep neural networks

Jian Wang , Yixing Yong , Haixia Bi , Lijun He , Fan Li

0 citations · 47 references · arXiv

α

Published on arXiv

2511.04970

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

Optimized geometric shapes (defined purely by their contour, without texture) can generate high-confidence misclassifications in DNNs and deceive downstream visual tasks, establishing a geometry-based adversarial paradigm distinct from pixel-level perturbation methods.

Fourier Shape Adversarial Framework

Novel technique introduced


While both shape and texture are fundamental to visual recognition, research on deep neural networks (DNNs) has predominantly focused on the latter, leaving their geometric understanding poorly probed. Here, we show: first, that optimized shapes can act as potent semantic carriers, generating high-confidence classifications from inputs defined purely by their geometry; second, that they are high-fidelity interpretability tools that precisely isolate a model's salient regions; and third, that they constitute a new, generalizable adversarial paradigm capable of deceiving downstream visual tasks. This is achieved through an end-to-end differentiable framework that unifies a powerful Fourier series to parameterize arbitrary shapes, a winding number-based mapping to translate them into the pixel grid required by DNNs, and signal energy constraints that enhance optimization efficiency while ensuring physically plausible shapes. Our work provides a versatile framework for probing the geometric world of DNNs and opens new frontiers for challenging and understanding machine perception.


Key Contributions

  • End-to-end differentiable framework using Fourier series to parameterize arbitrary closed shapes as adversarial inputs to DNNs
  • Winding number-based differentiable mapping that translates Fourier shape coefficients into pixel-grid images compatible with standard vision models
  • Signal energy regularization that constrains high-frequency components to produce physically plausible adversarial shapes while improving optimization efficiency

🛡️ Threat Analysis

Input Manipulation Attack

The paper's primary security contribution is a novel adversarial input paradigm — optimized closed contour shapes parameterized via Fourier series — that causes high-confidence misclassification in DNNs at inference time. The framework is end-to-end differentiable (gradient-based) and explicitly described as 'a new, generalizable adversarial paradigm capable of deceiving downstream visual tasks,' squarely within input manipulation attacks.


Details

Domains
vision
Model Types
cnntransformer
Threat Tags
white_boxinference_timetargeteddigital
Applications
image classificationobject detection