Brain-DiT: A Universal Multi-state fMRI Foundation Model with Metadata-Conditioned Pretraining
arXiv cs.CV / 4/15/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- Brain-DiT is proposed as a universal multi-state fMRI foundation model trained on 349,898 sessions across 24 datasets covering resting, task, naturalistic, disease, and sleep brain states.
- The model uses metadata-conditioned diffusion pretraining with a Diffusion Transformer (DiT), aiming to learn both fine-grained functional details and global semantic representations.
- Experiments across seven downstream tasks and multiple ablations suggest diffusion-based generative pretraining is a stronger proxy than masked reconstruction in raw/latent spaces or reconstruction/alignment approaches.
- Metadata conditioning is reported to improve downstream performance by disentangling intrinsic neural dynamics from population-level variability.
- The paper finds downstream objectives prefer different representational scales, with ADNI classification benefiting more from global semantics while age/sex prediction leans more on fine-grained local structure.
Related Articles
Are gamers being used as free labeling labor? The rise of "Simulators" that look like AI training grounds [D]
Reddit r/MachineLearning

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to
Failure to Reproduce Modern Paper Claims [D]
Reddit r/MachineLearning
Why don’t they just use Mythos to fix all the bugs in Claude Code?
Reddit r/LocalLLaMA