Tianrong Liu

h-index: 1 1 citations 1 papers (total)

Papers in Database (1)

benchmark arXiv Nov 9, 2025 · Nov 2025

Efficient LLM Safety Evaluation through Multi-Agent Debate

Dachuan Lin, Guobin Shen, Zihao Yang et al. · Beijing Institute of AI Safety and Governance · Chinese Academy of Sciences +3 more

Proposes SLM multi-agent debate judge and HAJailBench to evaluate LLM jailbreak safety at 43% lower inference cost

Prompt Injection nlp
1 citations PDF