Linking spatial biology and clinical histology via Haiku
arXiv cs.LG / 5/5/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces “Haiku,” a tri-modal contrastive learning model designed to jointly model spatial proteomics from multiplexed immunofluorescence alongside matched H&E histology and clinical metadata in a shared embedding space.
- Haiku is trained on a large, multi-patient dataset (26.7M spatial proteomics patches from 3,218 tissue sections across 1,606 patients, 11 organ types) and enables three-way cross-modal retrieval between molecular, histology, and clinical information.
- In multiple downstream tasks, Haiku improves over unimodal baselines, including cross-modal retrieval (Recall@50 up to 0.611), survival prediction (C-index 0.737, +7.91% relative), and zero-shot biomarker inference (mean Pearson correlation 0.718 over 52 biomarkers).
- The authors also propose a counterfactual prediction framework that changes only clinical metadata while holding tissue morphology fixed, surfacing niche-specific molecular shifts linked to breast cancer stage and lung cancer survival, presented explicitly as exploratory, hypothesis-generating signals.
- Overall, the work argues that tri-modal alignment can bridge spatial biology and clinical context to support integrative analysis and data-driven biological discovery rather than purely mechanistic conclusions.
Related Articles

Backed by Y Combinator and 20 unicorn founders, Moritz lands $9M
Tech.eu

Why Retail Chargeback Recovery Could Be AgentHansa's First Real PMF
Dev.to

Anthropic Launches AI Services Company with Blackstone & Goldman Sachs
Dev.to

Why B2B Revenue-Recovery Casework Looks Like AgentHansa's Best Early PMF
Dev.to

10 Ways AI Has Become Your Invisible Daily Companion in 2026
Dev.to