defense 2025

CluCERT: Certifying LLM Robustness via Clustering-Guided Denoising Smoothing

Zixia Wang , Gaojie Jin , Jia Hu , Ronghui Mu

0 citations · 41 references · arXiv

α

Published on arXiv

2512.08967

Input Manipulation Attack

OWASP ML Top 10 — ML01

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

CluCERT achieves tighter certified robustness bounds and lower computational cost than existing certified LLM robustness methods under synonym substitution attacks across multiple downstream tasks and jailbreak scenarios.

CluCERT

Novel technique introduced


Recent advancements in Large Language Models (LLMs) have led to their widespread adoption in daily applications. Despite their impressive capabilities, they remain vulnerable to adversarial attacks, as even minor meaning-preserving changes such as synonym substitutions can lead to incorrect predictions. As a result, certifying the robustness of LLMs against such adversarial prompts is of vital importance. Existing approaches focused on word deletion or simple denoising strategies to achieve robustness certification. However, these methods face two critical limitations: (1) they yield loose robustness bounds due to the lack of semantic validation for perturbed outputs and (2) they suffer from high computational costs due to repeated sampling. To address these limitations, we propose CluCERT, a novel framework for certifying LLM robustness via clustering-guided denoising smoothing. Specifically, to achieve tighter certified bounds, we introduce a semantic clustering filter that reduces noisy samples and retains meaningful perturbations, supported by theoretical analysis. Furthermore, we enhance computational efficiency through two mechanisms: a refine module that extracts core semantics, and a fast synonym substitution strategy that accelerates the denoising process. Finally, we conduct extensive experiments on various downstream tasks and jailbreak defense scenarios. Experimental results demonstrate that our method outperforms existing certified approaches in both robustness bounds and computational efficiency.


Key Contributions

  • Semantic clustering filter that retains meaningful perturbations and tightens certified robustness bounds, with formal theoretical analysis via a recovery factor γ
  • Refine module and fast synonym substitution strategy that reduce repeated sampling costs and improve computational efficiency
  • Empirical validation on multiple downstream NLP tasks and jailbreak defense scenarios, outperforming prior certified approaches on both bound tightness and efficiency

🛡️ Threat Analysis

Input Manipulation Attack

The paper's core contribution is certified robustness — a defense type explicitly listed under ML01 — providing provable guarantees against adversarial input perturbations (word substitutions) at inference time using an adaptation of randomized smoothing.


Details

Domains
nlp
Model Types
llmtransformer
Threat Tags
inference_timeblack_box
Applications
text classificationjailbreak defensellm safety