SCALE:Scalable Conditional Atlas-Level Endpoint transport for virtual cell perturbation prediction
arXiv cs.LG / 3/19/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- SCALE is a specialized large-scale foundation model for virtual cell perturbation prediction that jointly tackles training/inference bottlenecks and evaluation fidelity.
- It introduces a BioNeMo-based training and inference framework that achieves about 12.51x speedup in pretraining and 1.29x in inference compared with the prior state of the art under matched system settings.
- The perturbation prediction task is formulated as conditional transport using a set-aware flow that links LLaMA-based cellular encoding to endpoint-oriented supervision, improving training stability and perturbation recovery.
- Evaluation on Tahoe-100M with biologically meaningful metrics shows improvements: PDCorr up by 12.02% and DE Overlap up by 10.66% over STATE.
- The work argues for co-design of scalable infrastructure, stable transport modeling, and biologically faithful evaluation as essential for advancing virtual cell modeling.
Related Articles
The massive shift toward edge computing and local processing
Dev.to
Self-Refining Agents in Spec-Driven Development
Dev.to
The Three-Agent Protocol Is Transferable. The Discipline Isn't.
Dev.to
I built an abuse database for AI agents. It's free and open.
Dev.to

has anyone tried this? Flash-MoE: Running a 397B Parameter Model on a Laptop
Reddit r/LocalLLaMA