Lost in State Space: Probing Frozen Mamba Representations
arXiv cs.CL / 5/4/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper tests a hypothesis that Mamba’s recurrent state h_t can be used to extract token-level outputs at fixed patch boundaries to obtain semantic sentence representations without pooling heads or fine-tuning.
- Experiments on five benchmarks (SST-2, CoLA, MRPC, STS-B, IMDb) show that patch-boundary readouts from a frozen Mamba-130M backbone do not reliably beat simple mean pooling.
- The authors find major representation issues, including extremely high anisotropy and representational collapse in the raw final SSM state (e.g., MCC = 0.000 on CoLA across multiple seeds).
- They propose “orthogonal injection,” a modified recurrence intended to constrain how new information is incorporated, addressing the identified structural pathologies.
Related Articles
AnnouncementsBuilding a new enterprise AI services company with Blackstone, Hellman & Friedman, and Goldman Sachs
Anthropic News

Dara Khosrowshahi on replacing Uber drivers — and himself — with AI
The Verge

CLMA Frame Test
Dev.to

You Are Right — You Don't Need CLAUDE.md
Dev.to

Governance and Liability in AI Agents: What I Built Trying to Answer Those Questions
Dev.to