CourtGuard: A Model-Agnostic Framework for Zero-Shot Policy Adaptation in LLM Safety
Umid Suleymanov 1, Rufiz Bayramov 2, Suad Gafarli 2, Seljan Musayeva 2, Taghi Mammadov 2, Aynur Akhundlu 2, Murat Kantarcioglu 1
Published on arXiv
2602.22557
Prompt Injection
OWASP LLM Top 10 — LLM01
Key Finding
Outperforms fine-tuned policy-following baselines across 7 safety benchmarks and achieves 90% accuracy on an out-of-domain Wikipedia Vandalism task via zero-shot policy swapping
CourtGuard (Evidentiary Debate)
Novel technique introduced
Current safety mechanisms for Large Language Models (LLMs) rely heavily on static, fine-tuned classifiers that suffer from adaptation rigidity, the inability to enforce new governance rules without expensive retraining. To address this, we introduce CourtGuard, a retrieval-augmented multi-agent framework that reimagines safety evaluation as Evidentiary Debate. By orchestrating an adversarial debate grounded in external policy documents, CourtGuard achieves state-of-the-art performance across 7 safety benchmarks, outperforming dedicated policy-following baselines without fine-tuning. Beyond standard metrics, we highlight two critical capabilities: (1) Zero-Shot Adaptability, where our framework successfully generalized to an out-of-domain Wikipedia Vandalism task (achieving 90\% accuracy) by swapping the reference policy; and (2) Automated Data Curation and Auditing, where we leveraged CourtGuard to curate and audit nine novel datasets of sophisticated adversarial attacks. Our results demonstrate that decoupling safety logic from model weights offers a robust, interpretable, and adaptable path for meeting current and future regulatory requirements in AI governance.
Key Contributions
- CourtGuard: a retrieval-augmented multi-agent framework that frames LLM safety as Evidentiary Debate grounded in external policy documents, achieving SOTA on 7 safety benchmarks without fine-tuning
- Zero-shot policy adaptability via document swapping, demonstrated by generalizing to out-of-domain Wikipedia Vandalism detection at 90% accuracy
- Automated curation and auditing of nine novel adversarial attack datasets using the framework itself