CAGMamba: Context-Aware Gated Cross-Modal Mamba Network for Multimodal Sentiment Analysis

arXiv cs.CL / 4/7/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces CAGMamba, a context-aware gated cross-modal Mamba network designed for dialogue-based multimodal sentiment analysis (text + audio).
  • Instead of Transformer cross-modal attention with quadratic complexity, CAGMamba uses a Mamba-based design that provides explicit temporal structure by converting contextual and current-utterance features into a temporally ordered binary sequence.
  • It adds a Gated Cross-Modal Mamba Network (GCMN) that combines cross-modal and unimodal processing through learnable gating to better balance fusion quality with modality preservation.
  • The model is trained with a three-branch multi-task objective across text, audio, and fused predictions, improving sentiment evolution modeling across dialogue turns.
  • Experiments on three benchmark datasets show state-of-the-art or competitive performance across multiple metrics, and the authors provide code via a GitHub repository.

Abstract

Multimodal Sentiment Analysis (MSA) requires effective modeling of cross-modal interactions and contextual dependencies while remaining computationally efficient. Existing fusion approaches predominantly rely on Transformer-based cross-modal attention, which incurs quadratic complexity with respect to sequence length and limits scalability. Moreover, contextual information from preceding utterances is often incorporated through concatenation or independent fusion, without explicit temporal modeling that captures sentiment evolution across dialogue turns. To address these limitations, we propose CAGMamba, a context-aware gated cross-modal Mamba framework for dialogue-based sentiment analysis. Specifically, we organize the contextual and the current-utterance features into a temporally ordered binary sequence, which provides Mamba with explicit temporal structure for modeling sentiment evolution. To further enable controllable cross-modal integration, we propose a Gated Cross-Modal Mamba Network (GCMN) that integrates cross-modal and unimodal paths via learnable gating to balance information fusion and modality preservation, and is trained with a three-branch multi-task objective over text, audio, and fused predictions. Experiments on three benchmark datasets demonstrate that CAGMamba achieves state-of-the-art or competitive results across multiple evaluation metrics. All codes are available at https://github.com/User2024-xj/CAGMamba.