How Pruning Reshapes Features: Sparse Autoencoder Analysis of Weight-Pruned Language Models

arXiv cs.LG / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a systematic study of how unstructured weight pruning reshapes the internal feature geometry of language models, using Sparse Autoencoders (SAEs) as interpretability probes across multiple model families and sparsity levels.
  • It finds that rare SAE features (low firing rates) tend to survive pruning much better than frequent ones, suggesting pruning behaves like implicit feature selection that preferentially removes high-frequency generic features.
  • Wanda pruning is shown to preserve feature structure substantially better than magnitude pruning (up to about 3.7×), and SAE interpretability remains viable for Wanda-pruned models up to 50% sparsity.
  • The authors report a key dissociation: geometric survival of features under pruning does not reliably predict causal importance, highlighting limitations for using geometry alone to infer interpretability after compression.
  • The study examines stability, feature survival, SAE transferability, fragility, and causal relevance, providing multiple experimental insights relevant to interpreting compressed LLMs.

Abstract

Weight pruning is a standard technique for compressing large language models, yet its effect on learned internal representations remains poorly understood. We present the first systematic study of how unstructured pruning reshapes the feature geometry of language models, using Sparse Autoencoders (SAEs) as interpretability probes. Across three model families (Gemma 3 1B, Gemma 2 2B, Llama 3.2 1B), two pruning methods (magnitude and Wanda), and six sparsity levels (0--60%), we investigate five research questions spanning seed stability, feature survival, SAE transferability, feature fragility, and causal relevance. Our most striking finding is that rare SAE features--those with low firing rates--survive pruning far better than frequent ones, with within-condition Spearman correlations of rho = -1.0 in 11 of 17 experimental conditions. This counter-intuitive result suggests that pruning acts as implicit feature selection, preferentially destroying high-frequency generic features while preserving specialized rare ones. We further show that Wanda pruning preserves feature structure up to 3.7x better than magnitude pruning, that pre-trained SAEs remain viable on Wanda-pruned models up to 50% sparsity, and that geometric feature survival does not predict causal importance--a dissociation with implications for interpretability under compression.