RSGMamba: Reliability-Aware Self-Gated State Space Model for Multimodal Semantic Segmentation

arXiv cs.CV / 4/15/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes RSGMamba, a Reliability-aware Self-Gated State Space Model for multimodal semantic segmentation that treats cross-modal fusion as a modality-reliability problem rather than assuming all inputs are equally trustworthy.
  • Its core component, the Reliability-aware Self-Gated Mamba Block (RSGMB), explicitly models each modality’s reliability and uses self-gating to dynamically regulate how information is exchanged across modalities.
  • To improve fine-grained spatial detail, the method adds a lightweight Local Cross-Gated Modulation (LCGM) module that complements RSGMB’s global modeling.
  • Experiments report state-of-the-art results on RGB-D and RGB-T benchmarks, including 58.8%/54.0% mIoU on NYUDepth V2 and SUN-RGBD, and improvements up to +1.6% on MFNet and PST900 with 48.6M parameters.
  • Overall, the approach shows that reliability-aware feature selection and informative aggregation can mitigate degradation from noisy, misaligned, or incomplete auxiliary modalities.

Abstract

Multimodal semantic segmentation has emerged as a powerful paradigm for enhancing scene understanding by leveraging complementary information from multiple sensing modalities (e.g., RGB, depth, and thermal). However, existing cross-modal fusion methods often implicitly assume that all modalities are equally reliable, which can lead to feature degradation when auxiliary modalities are noisy, misaligned, or incomplete. In this paper, we revisit cross-modal fusion from the perspective of modality reliability and propose a novel framework termed the Reliability-aware Self-Gated State Space Model (RSGMamba). At the core of our method is the Reliability-aware Self-Gated Mamba Block (RSGMB), which explicitly models modality reliability and dynamically regulates cross-modal interactions through a self-gating mechanism. Unlike conventional fusion strategies that indiscriminately exchange information across modalities, RSGMB enables reliability-aware feature selection and enhancing informative feature aggregation. In addition, a lightweight Local Cross-Gated Modulation (LCGM) is incorporated to refine fine-grained spatial details, complementing the global modeling capability of RSGMB. Extensive experiments demonstrate that RSGMamba achieves state-of-the-art performance on both RGB-D and RGB-T semantic segmentation benchmarks, resulting 58.8% / 54.0% mIoU on NYUDepth V2 and SUN-RGBD (+0.4% / +0.7% over prior best), and 61.1% / 88.9% mIoU on MFNet and PST900 (up to +1.6%), with only 48.6M parameters, thereby validating the effectiveness and superiority of the proposed approach.