The Linear Centroids Hypothesis: How Deep Network Features Represent Data

arXiv cs.LG / 4/15/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes the Linear Centroids Hypothesis (LCH) to improve interpretability over the Linear Representation Hypothesis by characterizing features using linear directions of input-space centroids rather than latent activations alone.
  • LCH defines centroids as vector summaries of a deep network’s local functional behavior, aiming to avoid LRH’s limitations such as ignoring neuron/layer components and being vulnerable to spurious features.
  • The authors show that LCH-based interpretability can reuse existing LRH tooling (e.g., sparse autoencoders) by applying sparse feature learning to centroids instead of raw latent activations.
  • Experiments indicate that for DINO vision transformers, using centroids produces sparser feature dictionaries and also improves performance on downstream tasks.
  • The framework extends beyond vision models, suggesting that LCH can identify circuits in GPT-2 Large and includes released code to study the hypothesis.

Abstract

Identifying and understanding the features that a deep network (DN) extracts from its inputs to produce its outputs is a focal point of interpretability research. The Linear Representation Hypothesis (LRH) identifies features in terms of the linear directions formed by the inputs in a DN's latent space. However, the LRH is limited as it abstracts away from individual components (e.g., neurons and layers), is susceptible to identifying spurious features, and cannot be applied across sub-components (e.g., multiple layers). In this paper, we introduce the Linear Centroids Hypothesis (LCH) as a new framework for identifying the features of a DN. The LCH posits that features correspond to linear directions of centroids, which are vector summarizations of the functional behavior of a DN in a local region of its input space. Interpretability studies under the LCH can leverage existing LRH tools, such as sparse autoencoders, by applying them to the DN's centroids rather than to its latent activations. We demonstrate that doing so yields sparser feature dictionaries for DINO vision transformers, which also perform better on downstream tasks. The LCH also inspires novel approaches to interpretability; for example, LCH can readily identify circuits in GPT2-Large. For code to study the LCH https://github.com/ThomasWalker1/LinearCentroidsHypothesis .