Vision SmolMamba: Spike-Guided Token Pruning for Energy-Efficient Spiking State-Space Vision Models
arXiv cs.CV / 4/29/2026
📰 NewsModels & Research
Key Points
- The paper introduces Vision SmolMamba, a new spiking state-space vision architecture designed to improve efficiency over spiking Transformers that suffer from quadratic token interactions.
- It proposes a Spike-Guided Spatio-Temporal Token Pruner (SST-TP) that scores token importance using both spike activation strength and first-spike latency to remove redundant tokens progressively.
- The method integrates spike events directly into bidirectional state-space recurrence, yielding a spiking vision backbone intended for efficient long-range modeling.
- Experiments across static and event-based benchmarks (ImageNet-1K, CIFAR10/100, CIFAR10-DVS, DVS128 Gesture) show better accuracy-efficiency trade-offs than prior spiking Transformer baselines.
- The approach cuts estimated energy cost by at least 1.5× versus earlier spiking Transformer baselines and a Spiking Mamba variant while keeping competitive or improved accuracy.
Related Articles

How I Use AI Agents to Maintain a Living Knowledge Base for My Team
Dev.to
IK_LLAMA now supports Qwen3.5 MTP Support :O
Reddit r/LocalLLaMA
OpenAI models, Codex, and Managed Agents come to AWS
Dev.to

Automatic Error Recovery in AI Agent Networks
Dev.to
AeroJAX: JAX-native CFD, differentiable end-to-end. ~560 FPS at 128x128 on CPU [P]
Reddit r/MachineLearning