S3T-Former: A Purely Spike-Driven State-Space Topology Transformer for Skeleton Action Recognition
arXiv cs.AI / 3/20/2026
📰 NewsModels & Research
Key Points
- S3T-Former is proposed as the first purely spike-driven Transformer for energy-efficient skeleton action recognition, addressing power concerns on edge devices.
- It introduces Multi-Stream Anatomical Spiking Embedding (M-ASE) to convert multimodal skeleton features into highly sparse event streams, reducing dense computations.
- Lateral Spiking Topology Routing (LSTR) enables on-demand spike propagation and the Spiking State-Space (S3) Engine captures long-range temporal dynamics without non-sparse spectral processing.
- Experiments on large-scale datasets show competitive accuracy with theoretical energy savings, setting a new state-of-the-art in neuromorphic action recognition.
Related Articles

Math needs thinking time, everyday knowledge needs memory, and a new Transformer architecture aims to deliver both
THE DECODER
Kreuzberg v4.5.0: We loved Docling's model so much that we gave it a faster engine
Reddit r/LocalLLaMA
Today, what hardware to get for running large-ish local models like qwen 120b ?
Reddit r/LocalLLaMA
Running mistral locally for meeting notes and it's honestly good enough for my use case
Reddit r/LocalLLaMA
[D] Single-artist longitudinal fine art dataset spanning 5 decades now on Hugging Face — potential applications in style evolution, figure representation, and ethical training data
Reddit r/MachineLearning