defense 2025

ANCHOR: Integrating Adversarial Training with Hard-mined Supervised Contrastive Learning for Robust Representation Learning

Samarup Bhattacharya , Anubhab Bhattacharya , Abir Chakraborty

0 citations · 34 references · arXiv

α

Published on arXiv

2510.27599

Input Manipulation Attack

OWASP ML Top 10 — ML01

Key Finding

ANCHOR outperforms standard adversarial training methods on CIFAR-10 in both clean accuracy and robust accuracy under PGD-20 (ε=0.031), narrowing the accuracy-robustness trade-off gap.

ANCHOR

Novel technique introduced


Neural networks have changed the way machines interpret the world. At their core, they learn by following gradients, adjusting their parameters step by step until they identify the most discriminant patterns in the data. This process gives them their strength, yet it also opens the door to a hidden flaw. The very gradients that help a model learn can also be used to produce small, imperceptible tweaks that cause the model to completely alter its decision. Such tweaks are called adversarial attacks. These attacks exploit this vulnerability by adding tiny, imperceptible changes to images that, while leaving them identical to the human eye, cause the model to make wrong predictions. In this work, we propose Adversarially-trained Contrastive Hard-mining for Optimized Robustness (ANCHOR), a framework that leverages the power of supervised contrastive learning with explicit hard positive mining to enable the model to learn representations for images such that the embeddings for the images, their augmentations, and their perturbed versions cluster together in the embedding space along with those for other images of the same class while being separated from images of other classes. This alignment helps the model focus on stable, meaningful patterns rather than fragile gradient cues. On CIFAR-10, our approach achieves impressive results for both clean and robust accuracy under PGD-20 (epsilon = 0.031), outperforming standard adversarial training methods. Our results indicate that combining adversarial guidance with hard-mined contrastive supervision helps models learn more structured and robust representations, narrowing the gap between accuracy and robustness.


Key Contributions

  • ANCHOR framework that pairs supervised contrastive loss with adaptive hard positive mining, dynamically weighting intra-class samples most dissimilar to their class counterparts during adversarial training
  • Clustering adversarial perturbations, clean images, and augmentations together in embedding space to learn stable, non-fragile representations
  • Demonstrated improved clean and robust accuracy on CIFAR-10 under PGD-20 (ε=0.031) over standard adversarial training baselines

🛡️ Threat Analysis

Input Manipulation Attack

The paper proposes ANCHOR, a defense against adversarial input perturbations (FGSM, PGD, C&W attacks) that cause misclassification at inference time. The defense uses adversarial training augmented with hard-mined supervised contrastive learning to build representations that cluster adversarial examples with their clean counterparts, directly targeting adversarial robustness.


Details

Domains
vision
Model Types
cnn
Threat Tags
white_boxblack_boxinference_timetraining_timeuntargeteddigital
Datasets
CIFAR-10
Applications
image classification