A Comparative Study of Adversarial Robustness in CNN and CNN-ANFIS Architectures
Kaaustaaub Shankar , Bharadwaj Dogga , Kelly Cohen
Published on arXiv
2602.07028
Input Manipulation Attack
OWASP ML Top 10 — ML01
Key Finding
ANFIS integration is not universally beneficial for adversarial robustness: ResNet18-ANFIS gains robustness under PGD and Square attacks while VGG-ANFIS degrades relative to the VGG baseline.
DCNFIS
Novel technique introduced
Convolutional Neural Networks (CNNs) achieve strong image classification performance but lack interpretability and are vulnerable to adversarial attacks. Neuro-fuzzy hybrids such as DCNFIS replace fully connected CNN classifiers with Adaptive Neuro-Fuzzy Inference Systems (ANFIS) to improve interpretability, yet their robustness remains underexplored. This work compares standard CNNs (ConvNet, VGG, ResNet18) with their ANFIS-augmented counterparts on MNIST, Fashion-MNIST, CIFAR-10, and CIFAR-100 under gradient-based (PGD) and gradient-free (Square) attacks. Results show that ANFIS integration does not consistently improve clean accuracy and has architecture-dependent effects on robustness: ResNet18-ANFIS exhibits improved adversarial robustness, while VGG-ANFIS often underperforms its baseline. These findings suggest that neuro-fuzzy augmentation can enhance robustness in specific architectures but is not universally beneficial.
Key Contributions
- First systematic adversarial robustness comparison of standard CNNs (ConvNet, VGG, ResNet18) vs. their ANFIS-augmented (DCNFIS) counterparts across four datasets
- Demonstrates that neuro-fuzzy augmentation has architecture-dependent robustness effects: ResNet18-ANFIS improves robustness while VGG-ANFIS often underperforms its CNN baseline
- Evaluates both gradient-based (PGD) and gradient-free (Square) attacks to cover white-box and black-box threat models simultaneously
🛡️ Threat Analysis
The paper systematically evaluates CNN and CNN-ANFIS architectures under gradient-based (PGD, white-box) and gradient-free (Square, black-box) adversarial input manipulation attacks at inference time — adversarial robustness is the paper's central contribution.