Quantifying the Risk of Transferred Black Box Attacks
Disesdi Susanna Cox 1, Niklas Bunzel 2,3,4
Published on arXiv
2511.05102
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
CKA similarity-guided surrogate model selection improves adversarial subspace coverage, enabling more reliable regression-based risk estimates for transferred black-box evasion attacks than naive approaches
CKA-based surrogate selection for adversarial risk quantification
Novel technique introduced
Neural networks have become pervasive across various applications, including security-related products. However, their widespread adoption has heightened concerns regarding vulnerability to adversarial attacks. With emerging regulations and standards emphasizing security, organizations must reliably quantify risks associated with these attacks, particularly regarding transferred adversarial attacks, which remain challenging to evaluate accurately. This paper investigates the complexities involved in resilience testing against transferred adversarial attacks. Our analysis specifically addresses black-box evasion attacks, highlighting transfer-based attacks due to their practical significance and typically high transferability between neural network models. We underline the computational infeasibility of exhaustively exploring high-dimensional input spaces to achieve complete test coverage. As a result, comprehensive adversarial risk mapping is deemed impractical. To mitigate this limitation, we propose a targeted resilience testing framework that employs surrogate models strategically selected based on Centered Kernel Alignment (CKA) similarity. By leveraging surrogate models exhibiting both high and low CKA similarities relative to the target model, the proposed approach seeks to optimize coverage of adversarial subspaces. Risk estimation is conducted using regression-based estimators, providing organizations with realistic and actionable risk quantification.
Key Contributions
- Demonstrates computational infeasibility of exhaustive adversarial coverage testing in high-dimensional input spaces, making comprehensive adversarial risk mapping impractical
- Proposes strategically selecting surrogate models based on Centered Kernel Alignment (CKA) similarity — using both high and low CKA surrogates — to optimize coverage of adversarial subspaces
- Introduces regression-based risk estimators to provide organizations with actionable, realistic quantification of transferred adversarial attack risk
🛡️ Threat Analysis
The paper's core subject is transfer-based black-box evasion attacks — adversarial inputs crafted on surrogate models that transfer to target models at inference time. The proposed framework directly evaluates and quantifies resilience against these input manipulation attacks.