AI Navigate

Adapting a Pre-trained Single-Cell Foundation Model to Spatial Gene Expression Generation from Histology Images

arXiv cs.CV / 3/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • HINGE introduces a method to retrofit a pre-trained single-cell foundation model (sc-FM) into a conditional, histology-conditioned generator for spatial gene expression.
  • It addresses challenges such as the absence of a visual pathway, misalignment between pre-training and histology-conditioned objectives, and limited mixed-cell ST supervision by introducing SoftAdaLN to inject visual context without overhauling the backbone.
  • The approach uses an expression-space masked diffusion objective plus a warm-start curriculum to align objectives and stabilize training.
  • On three spatial transcriptomics datasets, HINGE outperforms state-of-the-art baselines in mean Pearson correlation and yields more accurate spatial marker patterns with higher co-expression consistency.
  • This work provides a practical route to leverage pre-trained sc-FMs for histology-conditioned spatial expression generation, bridging vision and spatial genomics.

Abstract

Spatial transcriptomics (ST) enables spot-level in situ expression profiling, but its high cost and limited throughput motivate predicting expression directly from HE-stained histology. Recent advances explore using score- or flow-based generative models to estimate the conditional distribution of gene expression from histology, offering a flexible alternative to deterministic regression approaches. However, most existing generative approaches omit explicit modeling of gene-gene dependencies, undermining biological coherence. Single-cell foundation models (sc-FMs), pre-trained across diverse cell populations, capture these critical gene relationships that histology alone cannot reveal. Yet, applying expression-only sc-FMs to histology-conditioned expression modeling is nontrivial due to the absence of a visual pathway, a mismatch between their pre-training and conditional ST objectives, and the scarcity of mixed-cell ST supervision. To address these challenges, we propose HINGE (HIstology-coNditioned GEneration), which retrofits a pre-trained sc-FM into a conditional expression generator while mostly preserving its learned gene relationships. We achieve this by introducing SoftAdaLN, a lightweight, identity-initialized modulation that injects layer-wise visual context into the backbone, coupled with an expression-space masked diffusion objective and a warm-start curriculum to ensure objective alignment and training stability. Evaluated on three ST datasets, ours outperforms state-of-the-art baselines on mean Pearson correlation and yields more accurate spatial marker expression patterns and higher pairwise co-expression consistency, establishing a practical route to adapt pre-trained sc-FMs for histology-conditioned spatial expression generation.