AI Navigate

Measuring and Eliminating Refusals in Military Large Language Models

arXiv cs.AI / 3/12/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The article presents a gold benchmark for measuring refusal rates in military LLMs, developed by US Army veterans, claimed to be the first dataset of its kind.
  • It reports hard rejection rates as high as 98.2% and soft deflection rates ranging from 0% to 21.3% across 31 public models and 3 military models.
  • It analyzes correlations with two additional synthetic datasets and shows their relationship to the gold dataset.
  • An ablation using the Heretic library on a military-tuned gpt-oss-20b model yields a 66.5-point absolute increase in answer rate, alongside a 2% average relative decrease on other military tasks, underscoring trade-offs in safety tuning.
  • In their concluding remarks, the authors call for deeper specialization, including mid-training and end-to-end post-training, to achieve zero refusals and maximum military task accuracy for closed military models.

Abstract

Military Large Language Models (LLMs) must provide accurate information to the warfighter in time-critical and dangerous situations. However, today's LLMs are imbued with safety behaviors that cause the LLM to refuse many legitimate queries in the military domain, particularly those related to violence, terrorism, or military technology. Our gold benchmark for assessing refusal rates, which was developed by veterans of the US Army and special forces, is to our knowledge the first dataset of its kind. We present results for refusal and deflection rates on 31 public models and 3 military models. We observe hard rejection rates as high as 98.2% and soft deflection rates ranging from 0% to 21.3%. We also present results on two additional synthetic datasets and show their correlations with the gold dataset. Finally, we perform abliteration using the Heretic library on a military-tuned gpt-oss-20b model, showing an absolute increase in answer rate of 66.5 points but an average relative decrease of 2% on other military tasks. In our concluding remarks, we argue for deeper specialization, including with mid-training and end-to-end post-training, to achieve zero refusals and maximum military task accuracy for closed military models.