Beyond Semantics: Disentangling Information Scope in Sparse Autoencoders for CLIP

arXiv cs.CV / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that interpreting Sparse Autoencoders (SAEs) for CLIP should go beyond feature-by-feature semantics by adding an “information scope” perspective.
  • It defines information scope as how broadly an SAE feature aggregates visual evidence, distinguishing localized, patch-specific cues from global, image-level signals.
  • The authors observe that some features remain stable under spatial perturbations while others change unpredictably with small input variations, suggesting fundamentally different scope behaviors.
  • They introduce the Contextual Dependency Score (CDS) to quantify this split between positionally stable local-scope features and positionally variant global-scope features.
  • Experimental results show that different scope types systematically affect CLIP predictions and confidence, making information scope a new diagnostic axis for SAE interpretability.

Abstract

Sparse Autoencoders (SAEs) have emerged as a powerful tool for interpreting the internal representations of CLIP vision encoders, yet existing analyses largely focus on the semantic meaning of individual features. We introduce information scope as a complementary dimension of interpretability that characterizes how broadly an SAE feature aggregates visual evidence, ranging from localized, patch-specific cues to global, image-level signals. We observe that some SAE features respond consistently across spatial perturbations, while others shift unpredictably with minor input changes, indicating a fundamental distinction in their underlying scope. To quantify this, we propose the Contextual Dependency Score (CDS), which separates positionally stable local scope features from positionally variant global scope features. Our experiments show that features of different information scopes exert systematically different influences on CLIP's predictions and confidence. These findings establish information scope as a critical new axis for understanding CLIP representations and provide a deeper diagnostic view of SAE-derived features.