Graph Signal Processing Meets Mamba2: Adaptive Filter Bank via Delta Modulation

arXiv cs.AI / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces HADES, a Graph Signal Processing (GSP)-inspired reinterpretation of the Mamba2 state-space model as an adaptive filter bank over a line graph.
  • HADES adds a hierarchical filter design using shared filters for global low-pass trends and expert filters for local high-pass behavior, controlled via structured bias on the delta parameter (Δ).
  • The authors report that HADES matches baseline performance (including Mamba2) on multiple tasks such as language modeling, commonsense reasoning, and long-context retrieval.
  • HADES reportedly achieves this with only 58.9% of the original parameters, aiming to improve efficiency without sacrificing benchmark quality.
  • The work claims to bridge GSP and neural sequence modeling by providing a more efficient and interpretable filtering perspective within SSM-based architectures.

Abstract

State-space models (SSMs) offer efficient alternatives to attention with linear-time recurrence. Mamba2, a recent SSM-based language model, uses selective input gating and a multi-head structure, enabling parallel computation and strong benchmark performance. However, its multi-head recurrence operates independently without structured utilization or analysis. In this work, we propose a novel method called Hierarchical ADaptive filter bank for Efficient SSMs (HADES), a Graph Signal Processing (GSP)-inspired framework that reinterprets Mamba2 as an adaptive filter bank on a line graph. Our hierarchical architecture introduces two filter types: shared filters for global low-pass behavior and expert filters for local high-pass behavior, achieved through structured bias on the parameter {\Delta}. HADES achieves comparable performance to baseline models including Mamba2 across various benchmarks in language modeling, commonsense reasoning, and long-context retrieval, while using only 58.9% of the original parameters. In this regard, HADES bridges GSP and neural sequence modeling, enabling efficient, hierarchical, and interpretable filtering within state-space models.