AI Navigate

Superficial Safety Alignment Hypothesis

arXiv cs.CL / 3/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The SSAH argues that safety alignment works as an implicit binary classifier guiding LLMs to either fulfill or refuse user requests, highlighting the distinct nature of safety alignment from general instruction-following.
  • The authors identify four attribute-critical components—Safety Critical Unit (SCU), Utility Critical Unit (UCU), Complex Unit (CU), and Redundant Unit (RU)—and define their roles in enforcing safety behavior.
  • They show that freezing certain safety-critical components during fine-tuning can preserve safety while allowing the model to adapt to new tasks, and that using redundant units as an "alignment budget" can reduce the alignment tax.
  • The paper argues that the atomic unit of safety lies at the neuron level, suggesting safety alignment need not be overly complex, and it provides code and project resources on its project site.
  • The work implies practical implications for developing safer LLMs, pointing to lightweight yet effective safety mechanisms that can be integrated with existing models.

Abstract

As large language models (LLMs) are overwhelmingly more and more integrated into various applications, ensuring they generate safe responses is a pressing need. Previous studies on alignment have largely focused on general instruction-following but have often overlooked the distinct properties of safety alignment, such as the brittleness of safety mechanisms. To bridge the gap, we propose the Superficial Safety Alignment Hypothesis (SSAH), which posits that safety alignment teaches an otherwise unsafe model to choose the correct reasoning direction-fulfill or refuse users' requests-interpreted as an implicit binary classification task. Through SSAH, we hypothesize that only a few essential components can establish safety guardrails in LLMs. We successfully identify four types of attribute-critical components: Safety Critical Unit (SCU), Utility Critical Unit (UCU), Complex Unit (CU), and Redundant Unit (RU). Our findings show that freezing certain safety-critical components during fine-tuning allows the model to retain its safety attributes while adapting to new tasks. Similarly, we show that leveraging redundant units in the pre-trained model as an "alignment budget" can effectively minimize the alignment tax while achieving the alignment goal. All considered, this paper concludes that the atomic functional unit for safety in LLMs is at the neuron level and underscores that safety alignment should not be complicated. We have code implementation and other information on the project website: https://ssa-h.github.io/.