Adapting a Pre-trained Single-Cell Foundation Model to Spatial Gene Expression Generation from Histology Images
arXiv cs.CV / 3/23/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- HINGE introduces a method to retrofit a pre-trained single-cell foundation model (sc-FM) into a conditional, histology-conditioned generator for spatial gene expression.
- It addresses challenges such as the absence of a visual pathway, misalignment between pre-training and histology-conditioned objectives, and limited mixed-cell ST supervision by introducing SoftAdaLN to inject visual context without overhauling the backbone.
- The approach uses an expression-space masked diffusion objective plus a warm-start curriculum to align objectives and stabilize training.
- On three spatial transcriptomics datasets, HINGE outperforms state-of-the-art baselines in mean Pearson correlation and yields more accurate spatial marker patterns with higher co-expression consistency.
- This work provides a practical route to leverage pre-trained sc-FMs for histology-conditioned spatial expression generation, bridging vision and spatial genomics.
Related Articles
Is AI becoming a bubble, and could it end like the dot-com crash?
Reddit r/artificial

Externalizing State
Dev.to

I made a 'benchmark' where LLMs write code controlling units in a 1v1 RTS game.
Dev.to

My AI Does Not Have a Clock
Dev.to
How to settle on a coding LLM ? What parameters to watch out for ?
Reddit r/LocalLLaMA