Supernodes and Halos: Loss-Critical Hubs in LLM Feed-Forward Layers

arXiv cs.LG / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper analyzes transformer feed-forward networks (FFNs) and shows that loss sensitivity is concentrated in a small fraction of channels, using a Fisher-style loss proxy based on activation-gradient second moments.
  • In Llama-3.1-8B, the top 1% of channels per layer accounts for a median of 58.7% of the loss-proxy (LP) mass, and the authors term these channels “supernodes.”
  • “Supernodes” only weakly overlap with activation-defined outliers and are not explained solely by activation power or weight norms, indicating a distinct loss-critical structure.
  • Beyond the supernode core, the authors observe a weaker “halo” where some non-supernode channels share write support and exhibit redundancy with the protected core.
  • One-shot structured FFN pruning experiments show that protecting supernodes (SCAR-Prot) preserves performance far better than baselines (perplexity 54.8 vs 989.2 at 50% sparsity), with similar LP-concentration patterns across multiple LLM families and scaling behavior during pretraining.

Abstract

We study the organization of channel-level importance in transformer feed-forward networks (FFNs). Using a Fisher-style loss proxy (LP) based on activation-gradient second moments, we show that loss sensitivity is concentrated in a small set of channels within each layer. In Llama-3.1-8B, the top 1% of channels per layer accounts for a median of 58.7% of LP mass, with a range of 33.0% to 86.1%. We call these loss-critical channels supernodes. Although FFN layers also contain strong activation outliers, LP-defined supernodes overlap only weakly with activation-defined outliers and are not explained by activation power or weight norms alone. Around this core, we find a weaker but consistent halo structure: some non-supernode channels share the supernodes' write support and show stronger redundancy with the protected core. We use one-shot structured FFN pruning as a diagnostic test of this organization. At 50% FFN sparsity, baselines that prune many supernodes degrade sharply, whereas our SCAR variants explicitly protect the supernode core; the strongest variant, SCAR-Prot, reaches perplexity 54.8 compared with 989.2 for Wanda-channel. The LP-concentration pattern appears across Mistral-7B, Llama-2-7B, and Qwen2-7B, remains visible in targeted Llama-3.1-70B experiments, and increases during OLMo-2-7B pretraining. These results suggest that LLM FFNs develop a small learned core of loss-critical channels, and that preserving this core is important for reliable structured pruning.