LEMON: a foundation model for nuclear morphology in Computational Pathology

arXiv cs.CV / 3/30/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • LEMON is introduced as a self-supervised foundation model designed to learn scalable, single-cell morphological representations for computational pathology.
  • The model is trained on millions of nucleus images spanning diverse tissues and cancer types, aiming to produce robust embeddings for cell-level analysis.
  • Researchers evaluate LEMON across five benchmark datasets and multiple prediction tasks, reporting strong performance and suggesting it can support a new paradigm for single-cell computational pathology.
  • Model weights are released publicly on Hugging Face, enabling other teams to reuse and build on LEMON for downstream pathology workflows.

Abstract

Computational pathology relies on effective representation learning to support cancer research and precision medicine. Although self-supervised learning has driven major progress at the patch and whole-slide image levels, representation learning at the single-cell level remains comparatively underexplored, despite its importance for characterizing cell types and cellular phenotypes. We introduce LEMON (Learning Embeddings from Morphology Of Nuclei), a self-supervised foundation model for scalable single-cell image representation learning. Trained on millions of cell images from diverse tissues and cancer types, LEMON learns robust and versatile morphological representations that support large-scale single-cell analyses in pathology. We evaluate LEMON on five benchmark datasets across a range of prediction tasks and show that it provides strong performance, highlighting its potential as a new paradigm for cell-level computational pathology. Model weights are available at https://huggingface.co/aliceblondel/LEMON.