Spatial Transcriptomics as Images for Large-Scale Pretraining
arXiv cs.CV / 3/17/2026
📰 NewsModels & Research
Key Points
- The paper addresses the ill-posed problem of defining a training sample for large-scale spatial transcriptomics pretraining, noting drawbacks of treating each spot as independent or an entire slide as a single sample.
- It proposes a croppable image-like representation by cropping fixed-size patches from raw slides to preserve spatial context while vastly increasing the number of training samples.
- The approach introduces gene-subset selection rules along the channel dimension to control input dimensionality and improve pretraining stability.
- Experiments show the image-like ST pretraining method consistently improves downstream performance over conventional schemes, with ablations confirming that both spatial patching and channel design are necessary.
Related Articles

報告:LLMにおける「自己言及的再帰」と「ステートフル・エミュレーション」の観測
note

諸葛亮 孔明老師(ChatGPTのロールプレイ)との対話 その肆拾伍『銀河文明・ダークマターエンジン』
note

GPT-5.4 mini/nano登場!―2倍高速で無料プランも使える小型高性能モデル
note

Why a Perfect-Memory AI Agent Without Persona Drift is Architecturally Impossible
Dev.to
OCP: Orthogonal Constrained Projection for Sparse Scaling in Industrial Commodity Recommendation
arXiv cs.LG