Beyond Semantics: Disentangling Information Scope in Sparse Autoencoders for CLIP
arXiv cs.CV / 4/8/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that interpreting Sparse Autoencoders (SAEs) for CLIP should go beyond feature-by-feature semantics by adding an “information scope” perspective.
- It defines information scope as how broadly an SAE feature aggregates visual evidence, distinguishing localized, patch-specific cues from global, image-level signals.
- The authors observe that some features remain stable under spatial perturbations while others change unpredictably with small input variations, suggesting fundamentally different scope behaviors.
- They introduce the Contextual Dependency Score (CDS) to quantify this split between positionally stable local-scope features and positionally variant global-scope features.
- Experimental results show that different scope types systematically affect CLIP predictions and confidence, making information scope a new diagnostic axis for SAE interpretability.
Related Articles
[N] Just found out that Milla Jovovich is a dev, invested in AI, and just open sourced a project
Reddit r/MachineLearning
ALTK‑Evolve: On‑the‑Job Learning for AI Agents
Hugging Face Blog
Context Windows Are Getting Absurd — And That's a Good Thing
Dev.to
Google isn’t an AI-first company despite Gemini being great
Reddit r/artificial
GitHub Weekly: Copilot SDK Goes Public, Cloud Agent Breaks Free
Dev.to