Sparse Autoencoders for Interpretable Medical Image Representation Learning

arXiv cs.CV / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study proposes Sparse Autoencoders (SAEs) to convert vision foundation model (FM) latent embeddings for medical images into human-interpretable, sparse features that clinicians can potentially interrogate and verify.
  • Trained on embeddings from BiomedParse and DINOv3 using 909,873 CT and MRI 2D slices from the TotalSegmentator dataset, the SAEs reconstruct original embeddings with high fidelity (R² up to 0.941) while retaining up to 87.8% of downstream performance using only 10 features (about 99.4% dimensionality reduction).
  • The learned sparse features maintain semantic fidelity for image retrieval tasks and show correspondence to specific concepts that can be described in language via LLM-based auto-interpretation.
  • The approach aims to bridge clinical language and abstract latent representations, enabling zero-shot, language-driven image retrieval through the SAE feature space.
  • The authors release a code repository, positioning SAEs as a promising route toward concept-driven and more interpretable medical vision systems.

Abstract

Vision foundation models (FMs) achieve state-of-the-art performance in medical imaging. However, they encode information in abstract latent representations that clinicians cannot interrogate or verify. The goal of this study is to investigate Sparse Autoencoders (SAEs) for replacing opaque FM image representations with human-interpretable, sparse features. We train SAEs on embeddings from BiomedParse (biomedical) and DINOv3 (general-purpose) using 909,873 CT and MRI 2D image slices from the TotalSegmentator dataset. We find that learned sparse features: (a) reconstruct original embeddings with high fidelity (R2 up to 0.941) and recover up to 87.8% of downstream performance using only 10 features (99.4% dimensionality reduction), (b) preserve semantic fidelity in image retrieval tasks, (c) correspond to specific concepts that can be expressed in language using large language model (LLM)-based auto-interpretation. (d) bridge clinical language and abstract latent representations in zero-shot language-driven image retrieval. Our work indicates SAEs are a promising pathway towards interpretable, concept-driven medical vision systems. Code repository: https://github.com/pwesp/sail.

Sparse Autoencoders for Interpretable Medical Image Representation Learning | AI Navigate