Diffusion-Based Feature Denoising with NNMF for Robust handwritten digit multi-class classification
Hiba Adil Al-kharsan , Róbert Rajkó
Published on arXiv
2603.29917
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
Diffusion-based hybrid model outperforms CNN baseline models in adversarial settings while maintaining competitive clean accuracy
Diffusion-Based Feature Denoising with NNMF
Novel technique introduced
This work presents a robust multi-class classification framework for handwritten digits that combines diffusion-driven feature denoising with a hybrid feature representation. Inspired by our previous work on brain tumor classification, the proposed approach operates in a feature space to improve the robustness to noise and adversarial attacks. First, the input images are converted into tight, interpretable exemplification using Nonnegative Matrix Factorization (NNMF). In parallel, special deep features are extracted using a computational neural network (CNN). These integral features are combined into a united hybrid representation. To improve robustness, a step diffusion operation is used in the feature space by gradually adding Gaussian noise. A feature denoiser network is trained to reverse this operation and rebuild clean representations from tilted inputs. The courteous features are then applied for multi-class classification. The suggested method is evaluated in both baseline and adversarial settings using AutoAttack. The experimental outcome present that the diffusion-based hybrid model is both effective and robust, the CNN baseline models outperforming while maintain powerful classification performance. These results explain the activity of feature-level diffusion defense for reliable multi-class handwritten digit classification.
Key Contributions
- Hybrid feature representation combining CNN deep features with NNMF interpretable features for robust classification
- Diffusion-based feature denoising mechanism operating in feature space rather than image space to reduce computational complexity
- Evaluation against AutoAttack framework demonstrating improved robustness while maintaining baseline performance
🛡️ Threat Analysis
The paper explicitly defends against adversarial examples (AutoAttack framework) that cause misclassification at inference time. The diffusion-based denoising operates in feature space to purify adversarial perturbations, which is a direct defense against input manipulation attacks.