ARMOR 2025: A Military-Aligned Benchmark for Evaluating Large Language Model Safety Beyond Civilian Contexts

arXiv cs.AI / 5/4/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes ARMOR 2025, a large-language-model safety benchmark designed for defense use cases that go beyond generic “civilian” social-risk testing.
  • ARMOR 2025 is grounded in three military doctrine sources—the Law of War, Rules of Engagement, and Joint Ethics Regulation—and uses doctrinal text to create meaning-preserving multiple-choice questions.
  • The benchmark is organized using an OODA (Observe–Orient–Decide–Act)-informed taxonomy to systematically test both accuracy and refusal behavior across military-relevant decision types.
  • In evaluations against 21 commercial LLMs, the authors found significant gaps in safety alignment for military decision-support scenarios.
  • The benchmark includes a structured 12-category taxonomy, 519 prompts, and rigorous evaluation procedures to enable more realistic assessment of legal/ethical compliance.

Abstract

Large language models (LLMs) are now being explored for defense applications that require reliable and legally compliant decision support. They also hold significant potential to enhance decision making, coordination, and operational efficiency in military contexts. These uses demand evaluation methods that reflect the doctrinal standards that guide real military operations. Existing safety benchmarks focus on general social risks and do not test whether models follow the legal and ethical rules that govern real military operations. To address this gap, we introduce ARMOR 2025, a military aligned safety benchmark grounded in three core military doctrines the Law of War, the Rules of Engagement, and the Joint Ethics Regulation. We extract doctrinal text from these sources and generate multiple choice questions that preserve the intended meaning of each rule. The benchmark is organized through a taxonomy informed by the Observe Orient Decide Act (OODA) decision making framework. This structure enables systematic testing of accuracy and refusal across military relevant decision types. This benchmark features a structured 12-category taxonomy, 519 doctrinally grounded prompts, and rigorous evaluation procedures applied to 21 commercial LLMs. Evaluation results reveal critical gaps in safety alignment for military applications.