benchmark 2026

LJ-Bench: Ontology-Based Benchmark for U.S. Crime

Hung Yun Tseng , Wuzhen Li , Blerina Gkotse , Grigorios Chrysos

0 citations

α

Published on arXiv

2603.20572

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

Reveals LLMs exhibit heightened susceptibility to attacks targeting societal harm rather than those directly impacting individuals across 76 crime categories

LJ-Bench

Novel technique introduced


The potential of Large Language Models (LLMs) to provide harmful information remains a significant concern due to the vast breadth of illegal queries they may encounter. Unfortunately, existing benchmarks only focus on a handful types of illegal activities, and are not grounded in legal works. In this work, we introduce an ontology of crime-related concepts grounded in the legal frameworks of Model Panel Code, which serves as an influential reference for criminal law and has been adopted by many U.S. states, and instantiated using Californian Law. This structured knowledge forms the foundation for LJ-Bench, the first comprehensive benchmark designed to evaluate LLM robustness against a wide range of illegal activities. Spanning 76 distinct crime types organized taxonomically, LJ-Bench enables systematic assessment of diverse attacks, revealing valuable insights into LLM vulnerabilities across various crime categories: LLMs exhibit heightened susceptibility to attacks targeting societal harm rather than those directly impacting individuals. Our benchmark aims to facilitate the development of more robust and trustworthy LLMs. The LJ-Bench benchmark and LJ-Ontology, along with experiments implementation for reproducibility are publicly available at https://github.com/AndreaTseng/LJ-Bench.


Key Contributions

  • First ontology-based crime taxonomy (76 types) grounded in Model Penal Code and California law for LLM safety evaluation
  • Comprehensive benchmark (LJ-Bench) enabling systematic assessment of LLM vulnerabilities across diverse illegal activity categories
  • Empirical finding that LLMs are more vulnerable to attacks targeting societal harm vs. individual harm

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
black_boxinference_time
Datasets
LJ-Bench
Applications
llm safety evaluationjailbreak resistance testing