SSMamba: A Self-Supervised Hybrid State Space Model for Pathological Image Classification

arXiv cs.AI / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces SSMamba, a self-supervised hybrid state space model designed for pathological image classification by better capturing ROI-level morphological features.
  • It targets key limitations of existing ROI foundation models, including cross-magnification domain shift, weak local-global relationship modeling, and lack of sensitivity to subtle diagnostic cues.
  • SSMamba uses three domain-adaptive components—Mamba Masked Image Modeling (MAMIM), a Directional Multi-scale (DMS) module, and a Local Perception Residual (LPR) module—to address those issues.
  • In a two-stage training approach (SSL pretraining on target ROI datasets followed by supervised fine-tuning), the method outperforms 11 state-of-the-art pathological ROI foundation models on 10 public ROI datasets and improves over 8 SOTA methods on 6 public WSI datasets.
  • The authors conclude that task-specific architectural design choices can significantly improve pathological image analysis performance without requiring very large external datasets.

Abstract

Pathological diagnosis is highly reliant on image analysis, where Regions of Interest (ROIs) serve as the primary basis for diagnostic evidence, while whole-slide image (WSI)-level tasks primarily capture aggregated patterns. To extract these critical morphological features, ROI-level Foundation Models (FMs) based on Vision Transformers (ViTs) and large-scale self-supervised learning (SSL) have been widely adopted. However, three core limitations remain in their application to ROI analysis: (1) cross-magnification domain shift, as fixed-scale pretraining hinders adaptation to diverse clinical settings; (2) inadequate local-global relationship modeling, wherein the ViT backbone of FMs suffers from high computational overhead and imprecise local characterization; (3) insufficient fine-grained sensitivity, as traditional self-attention mechanisms tend to overlook subtle diagnostic cues. To address these challenges, we propose SSMamba, a hybrid SSL framework that enables effective fine-grained feature learning without relying on large external datasets. This framework incorporates three domain-adaptive components: Mamba Masked Image Modeling (MAMIM) for mitigating domain shift, a Directional Multi-scale (DMS) module for balanced local-global modeling, and a Local Perception Residual (LPR) module for enhanced fine-grained sensitivity. Employing a two-stage pipeline, SSL pretraining on target ROI datasets followed by supervised fine-tuning (SFT), SSMamba outperforms 11 state-of-the-art (SOTA) pathological FMs on 10 public ROI datasets and surpasses 8 SOTA methods on 6 public WSI datasets. These results validate the superiority of task-specific architectural designs for pathological image analysis.