LJ-Bench: Ontology-Based Benchmark for U.S. Crime
arXiv cs.LG / 2026/3/24
📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research
要点
- The paper introduces LJ-Bench, a new benchmark that evaluates how robust large language models are against a broad set of illegal/crime-related queries.
- LJ-Bench is grounded in an ontology of crime concepts derived from the Model Penal Code and instantiated using California law, enabling legally grounded and structured coverage.
- The benchmark covers 76 distinct crime types arranged taxonomically, supporting systematic testing across many categories rather than a small set of isolated illegal activities.
- Experiments indicate LLMs are more vulnerable to attacks targeting societal harm than to those directly affecting individuals.
- The LJ-Bench benchmark, LJ-Ontology, and implementation code are released publicly to support reproducible research and the development of safer, more trustworthy LLMs.
