AI Navigate

Deactivating Refusal Triggers: Understanding and Mitigating Overrefusal in Safety Alignment

arXiv cs.AI / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates overrefusal in safety-aligned LLMs, showing that models can reject benign queries after safety alignment due to how refusal cues are learned.
  • It defines refusal triggers as linguistic cues in training data that elicit refusal responses, and explains how safety alignment can cause models to associate these triggers with refusal, leading to overrefusal.
  • The authors propose a mitigation strategy that explicitly accounts for refusal triggers during safety-alignment fine-tuning to balance defense against harmful inputs with responsiveness to benign queries.
  • Empirical results indicate the proposed method improves the trade-off between resisting jailbreaks and remaining helpful on benign queries, but the paper warns that it contains harmful and biased sentences.

Abstract

Safety alignment aims to ensure that large language models (LLMs) refuse harmful requests by post-training on harmful queries paired with refusal answers. Although safety alignment is widely adopted in industry, the overrefusal problem where aligned LLMs also reject benign queries after safety alignment post-training, remains insufficiently studied. Such an issue degrades the usability of safety alignment in real-world applications. In this paper, we examine how overrefusal arises under safety alignment, and propose a mitigation strategy inspired by our findings. We define refusal triggers as linguistic cues in the training data that elicit refusal responses, safety alignment encourages LLMs to associate refusal triggers within a training sample with refusal responses, leading aligned LLMs to refuse harmful queries. However, the refusal triggers include not only harmful linguistic cues but also non-harmful cues, therefore causing overrefusal to benign queries. Building on this mechanistic analysis, we propose a method that explicitly considers refusal triggers in the safety alignment fine-tuning. Empirical results demonstrate that our approach achieves a more favorable trade-off between defense against jailbreak attacks and responsiveness to benign queries, outperforming prior methods. Warning: this paper contains harmful and biased sentences.