attack 2025

Geometry-Aware Backdoor Attacks: Leveraging Curvature in Hyperbolic Embeddings

Ali Baheri

0 citations · 28 references · arXiv

α

Published on arXiv

2510.06397

Model Poisoning

OWASP ML Top 10 — ML10

Key Finding

Attack success rate increases monotonically toward the Poincaré disk boundary while standard detector sensitivity decreases, empirically confirming the theoretically predicted geometry-specific vulnerability in hyperbolic neural networks.

Geometry-Aware Backdoor Attack

Novel technique introduced


Non-Euclidean foundation models increasingly place representations in curved spaces such as hyperbolic geometry. We show that this geometry creates a boundary-driven asymmetry that backdoor triggers can exploit. Near the boundary, small input changes appear subtle to standard input-space detectors but produce disproportionately large shifts in the model's representation space. Our analysis formalizes this effect and also reveals a limitation for defenses: methods that act by pulling points inward along the radius can suppress such triggers, but only by sacrificing useful model sensitivity in that same direction. Building on these insights, we propose a simple geometry-adaptive trigger and evaluate it across tasks and architectures. Empirically, attack success increases toward the boundary, whereas conventional detectors weaken, mirroring the theoretical trends. Together, these results surface a geometry-specific vulnerability in non-Euclidean models and offer analysis-backed guidance for designing and understanding the limits of defenses.


Key Contributions

  • Formal theoretical analysis showing that the boundary of hyperbolic (Poincaré disk) space creates an asymmetry where small input perturbations produce disproportionately large representation shifts, enabling stealthy backdoor triggers
  • Geometry-adaptive backdoor trigger design that exploits this boundary-driven amplification to maximize attack success while remaining subtle to standard input-space detectors
  • Analysis of the inherent trade-off in radial-pulling defenses: suppressing boundary-region triggers requires sacrificing model sensitivity in the same direction, bounding defensibility

🛡️ Threat Analysis

Model Poisoning

Paper proposes geometry-adaptive backdoor triggers that exploit the Poincaré disk boundary effect to embed hidden, targeted malicious behavior in hyperbolic neural networks while evading standard input-space detectors — a direct backdoor/trojan attack contribution.


Details

Domains
graph
Model Types
gnn
Threat Tags
training_timetargeteddigital
Applications
knowledge graph embeddingrecommendation systemsfraud detectiondrug discoverysocial network analysis