Automated Analysis of Global AI Safety Initiatives: A Taxonomy-Driven LLM Approach

arXiv cs.AI / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces an automated “crosswalk” framework that compares two AI safety policy documents by extracting activities and mapping them to a shared taxonomy (Activity Map on AI Safety).
  • For each taxonomy aspect, the system generates short summaries, brief comparisons, and similarity scores, including heatmap visualizations of mean similarities across model runs.
  • Experiments using five large language models on ten public AI safety documents show that model choice strongly influences crosswalk results and can lead to high disagreement between documents across models.
  • Human evaluation by three experts finds strong inter-annotator agreement on two document pairs, but LLM-derived similarity scores still do not fully align with human judgments.
  • Overall, the study supports using LLM-based comparative inspection of policy documents while emphasizing the need to account for model-dependent variability and validation against human assessments.

Abstract

We present an automated crosswalk framework that compares an AI safety policy document pair under a shared taxonomy of activities. Using the activity categories defined in Activity Map on AI Safety as fixed aspects, the system extracts and maps relevant activities, then produces for each aspect a short summary for each document, a brief comparison, and a similarity score. We assess the stability and validity of LLM-based crosswalk analysis across public policy documents. Using five large language models, we perform crosswalks on ten publicly available documents and visualize mean similarity scores with a heatmap. The results show that model choice substantially affects the crosswalk outcomes, and that some document pairs yield high disagreements across models. A human evaluation by three experts on two document pairs shows high inter-annotator agreement, while model scores still differ from human judgments. These findings support comparative inspection of policy documents.

Automated Analysis of Global AI Safety Initiatives: A Taxonomy-Driven LLM Approach | AI Navigate