Can Natural Image Autoencoders Compactly Tokenize fMRI Volumes for Long-Range Dynamics Modeling?
arXiv cs.CV / 4/7/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes TABLeT, a Two-dimensionally Autoencoded Brain Latent Transformer that tokenizes 3D fMRI volumes into compact continuous tokens to make long-range spatiotemporal modeling feasible under limited memory.
- By leveraging a pre-trained 2D natural image autoencoder, each fMRI volume is compressed into tokens that can be processed by a simple Transformer encoder while reducing VRAM requirements compared with voxel-based approaches.
- Experiments on large benchmarks (UK-Biobank, HCP, and ADHD-200) show TABLeT outperforming existing models across multiple tasks.
- The authors introduce a self-supervised masked token modeling pre-training method for TABLeT, which further improves downstream performance.
- The work claims gains in computational and memory efficiency while aiming to preserve interpretability for scalable brain-activity dynamics modeling, with code released on GitHub.
Related Articles

Black Hat Asia
AI Business

Meta Superintelligence Lab Releases Muse Spark: A Multimodal Reasoning Model With Thought Compression and Parallel Agents
MarkTechPost

Chatbots are great at manipulating people to buy stuff, Princeton boffins find
The Register
I tested and ranked every ai companion app I tried and here's my honest breakdown
Reddit r/artificial

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to