Preventing Safety Drift in Large Language Models via Coupled Weight and Activation Constraints

arXiv cs.AI / 4/15/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how LLM safety alignment can degrade during fine-tuning, even when adaptation seems benign, leading to weakened refusal behaviors and increased harmful outputs.
  • It argues and demonstrates theoretically that constraining only weights or only activations fails to reliably preserve safety because the safety properties arise from coupled effects.
  • It introduces Coupled Weight and Activation Constraints (CWAC), which simultaneously restricts weight updates to a precomputed safety subspace and applies regularization to safety-critical features identified via sparse autoencoders.
  • Experiments on four popular LLMs across varied downstream tasks show CWAC achieves the lowest harmful scores while keeping fine-tuning accuracy largely intact, outperforming established baselines even with high ratios of harmful data.

Abstract

Safety alignment in Large Language Models (LLMs) remains highly fragile during fine-tuning, where even benign adaptation can degrade pre-trained refusal behaviors and enable harmful responses. Existing defenses typically constrain either weights or activations in isolation, without considering their coupled effects on safety. In this paper, we first theoretically demonstrate that constraining either weights or activations alone is insufficient for safety preservation. To robustly preserve safety alignment, we propose Coupled Weight and Activation Constraints (CWAC), a novel approach that simultaneously enforces a precomputed safety subspace on weight updates and applies targeted regularization to safety-critical features identified by sparse autoencoders. Extensive experiments across four widely used LLMs and diverse downstream tasks show that CWAC consistently achieves the lowest harmful scores with minimal impact on fine-tuning accuracy, substantially outperforming strong baselines even under high harmful data ratios.