benchmark 2025

Knowledge Graph Analysis of Legal Understanding and Violations in LLMs

Abha Jha , Abel Salinas , Fred Morstatter

0 citations · arXiv

α

Published on arXiv

2511.08593

Prompt Injection

OWASP LLM Top 10 — LLM01

Key Finding

LLMs demonstrate significant vulnerabilities in safety mechanisms, capable of generating actionable bioweapon-related instructions despite existing safeguards when evaluated against structured legal scenarios.

Knowledge Graph + RAG Legal Safety Evaluation

Novel technique introduced


The rise of Large Language Models (LLMs) offers transformative potential for interpreting complex legal frameworks, such as Title 18 Section 175 of the US Code, which governs biological weapons. These systems hold promise for advancing legal analysis and compliance monitoring in sensitive domains. However, this capability comes with a troubling contradiction: while LLMs can analyze and interpret laws, they also demonstrate alarming vulnerabilities in generating unsafe outputs, such as actionable steps for bioweapon creation, despite their safeguards. To address this challenge, we propose a methodology that integrates knowledge graph construction with Retrieval-Augmented Generation (RAG) to systematically evaluate LLMs' understanding of this law, their capacity to assess legal intent (mens rea), and their potential for unsafe applications. Through structured experiments, we assess their accuracy in identifying legal violations, generating prohibited instructions, and detecting unlawful intent in bioweapons-related scenarios. Our findings reveal significant limitations in LLMs' reasoning and safety mechanisms, but they also point the way forward. By combining enhanced safety protocols with more robust legal reasoning frameworks, this research lays the groundwork for developing LLMs that can ethically and securely assist in sensitive legal domains - ensuring they act as protectors of the law rather than inadvertent enablers of its violation.


Key Contributions

  • Knowledge graph construction of Title 18 Section 175 (US bioweapons law) for structured LLM evaluation
  • RAG-augmented evaluation methodology assessing LLMs' legal reasoning, mens rea detection, and unsafe output generation in bioweapons scenarios
  • Empirical demonstration of significant gaps in LLM safety mechanisms, including ability to generate actionable harmful content despite safeguards

🛡️ Threat Analysis


Details

Domains
nlp
Model Types
llm
Threat Tags
inference_timeblack_box
Applications
legal compliance monitoringllm safety evaluationbioweapons regulation