Critical-CoT: A Robust Defense Framework against Reasoning-Level Backdoor Attacks in Large Language Models
Published on arXiv
2604.10681
Model Poisoning
OWASP ML Top 10 — ML10
Key Finding
Achieves strong backdoor detection and suppression across multiple LLMs and datasets against both ICL-based and FT-based reasoning-level attacks
Critical-CoT
Novel technique introduced
Large Language Models (LLMs), despite their impressive capabilities across domains, have been shown to be vulnerable to backdoor attacks. Prior backdoor strategies predominantly operate at the token level, where an injected trigger causes the model to generate a specific target word, choice, or class (depending on the task). Recent advances, however, exploit the long-form reasoning tendencies of modern LLMs to conduct reasoning-level backdoors: once triggered, the victim model inserts one or more malicious reasoning steps into its chain-of-thought (CoT). These attacks are substantially harder to detect, as the backdoored answer remains plausible and consistent with the poisoned reasoning trajectory. Yet, defenses tailored to this type of backdoor remain largely unexplored. To bridge this gap, we propose Critical-CoT, a novel defense mechanism that conducts a two-stage fine-tuning (FT) process on LLMs to develop critical thinking behaviors, enabling them to automatically identify potential backdoors and refuse to generate malicious reasoning steps. Extensive experiments across multiple LLMs and datasets demonstrate that Critical-CoT provides strong robustness against both in-context learning-based and FT-based backdoor attacks. Notably, Critical-CoT exhibits strong cross-domain and cross-task generalization. Our code is available at hthttps://github.com/tuanvu171/Critical-CoT.
Key Contributions
- Novel two-stage fine-tuning defense (SFT + DPO) that teaches LLMs critical thinking to identify backdoor triggers
- First unified defense effective against both ICL-based and FT-based reasoning-level backdoor attacks
- Strong cross-domain and cross-task generalization without requiring prior knowledge of triggers or attack strategies
🛡️ Threat Analysis
Primary focus is defending against backdoor attacks that inject malicious reasoning steps into LLM chain-of-thought outputs via both fine-tuning and in-context learning poisoning.