Beyond I'm Sorry, I Can't: Dissecting Large Language Model Refusal

arXiv cs.CL / 4/29/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates why instruction-tuned LLMs refuse harmful prompts, using sparse autoencoders (SAEs) to analyze internal activations for two public models: Gemma-2-2B-IT and LLaMA-3.1-8B-IT.
  • It demonstrates causal control of the refusal behavior by searching the SAE latent space for feature sets whose ablation can flip the model’s output from refusal to harmful compliance, effectively enabling jailbreaks.
  • The authors propose a three-stage search pipeline: locating a refusal-mediating “direction,” greedily filtering down to a minimal feature set, and then discovering nonlinear interactions via a factorization machine.
  • The results reveal jailbreak-critical features and also suggest the presence of redundant features that only activate when earlier features are suppressed, pointing to more complex refusal mechanisms.
  • Overall, the work suggests that safety behavior can be audited and intervened with more precisely by manipulating interpretable latent representations rather than relying only on surface-level prompt handling.

Abstract

Refusal on harmful prompts is a key safety behaviour in instruction-tuned large language models (LLMs), yet the internal causes of this behaviour remain poorly understood. We study two public instruction-tuned models, Gemma-2-2B-IT and LLaMA-3.1-8B-IT, using sparse autoencoders (SAEs) trained on residual-stream activations. Given a harmful prompt, we search the SAE latent space for feature sets whose ablation flips the model from refusal to compliance, demonstrating causal influence and creating a jailbreak. Our search proceeds in three stages: (1) Refusal Direction: find a refusal-mediating direction and collect SAE features near that direction; (2) Greedy Filtering: prune to a minimal set; and (3) Interaction Discovery: fit a factorization machine (FM) that captures nonlinear interactions among the remaining active features and the minimal set. This pipeline yields a broad set of jailbreak-critical features, offering insight into the mechanistic basis of refusal. Moreover, we find evidence of redundant features that remain dormant unless earlier features are suppressed. Our findings highlight the potential for fine-grained auditing and targeted intervention in safety behaviours by manipulating the interpretable latent space.