LJ-Bench: Ontology-Based Benchmark for U.S. Crime
arXiv cs.LG / 3/24/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces LJ-Bench, a new benchmark that evaluates how robust large language models are against a broad set of illegal/crime-related queries.
- LJ-Bench is grounded in an ontology of crime concepts derived from the Model Penal Code and instantiated using California law, enabling legally grounded and structured coverage.
- The benchmark covers 76 distinct crime types arranged taxonomically, supporting systematic testing across many categories rather than a small set of isolated illegal activities.
- Experiments indicate LLMs are more vulnerable to attacks targeting societal harm than to those directly affecting individuals.
- The LJ-Bench benchmark, LJ-Ontology, and implementation code are released publicly to support reproducible research and the development of safer, more trustworthy LLMs.
Related Articles

Composer 2: What is new and Compares with Claude Opus 4.6 & GPT-5.4
Dev.to
How UCP Breaks Your E-Commerce Tracking Stack: A Platform-by-Platform Analysis
Dev.to
AI Text Analyzer vs Asking Friends: Which Gives Better Perspective?
Dev.to
[D] Cathie wood claims ai productivity wave is starting, data shows 43% of ceos save 8+ hours weekly
Reddit r/MachineLearning

Microsoft hires top AI researchers from Allen Institute for AI for Suleyman's Superintelligence team
THE DECODER