Ultrasound-CLIP: Semantic-Aware Contrastive Pre-training for Ultrasound Image-Text Understanding

arXiv cs.CV / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Ultrasound-CLIP, a semantic-aware contrastive pretraining method tailored to ultrasound image–text understanding rather than directly reusing CLIP-style models designed for other modalities.
  • It builds the US-365K dataset with 365k paired ultrasound images and text labels across 52 anatomical categories, alongside a structured knowledge system using an Ultrasonographic Hierarchical Anatomical Taxonomy (UDT) and a nine-dimension Diagnostic Attribute Framework (UDAF).
  • Ultrasound-CLIP improves contrastive learning via semantic soft labels and a semantic loss to better refine discrimination among heterogeneous ultrasound samples.
  • The approach also constructs a heterogeneous graph modality from UDAF-derived text representations to enable structured reasoning over lesion–attribute relationships.
  • Experiments using patient-level splits show state-of-the-art results on classification and retrieval, with strong generalization across zero-shot, linear probing, and fine-tuning settings.

Abstract

Ultrasound imaging is widely used in clinical diagnostics due to its real-time capability and radiation-free nature. However, existing vision-language pre-training models, such as CLIP, are primarily designed for other modalities, and are difficult to directly apply to ultrasound data, which exhibit heterogeneous anatomical structures and diverse diagnostic attributes. To bridge this gap, we construct US-365K, a large-scale ultrasound image-text dataset containing 365k paired samples across 52 anatomical categories. We establish Ultrasonographic Diagnostic Taxonomy (UDT) containing two hierarchical knowledge frameworks. Ultrasonographic Hierarchical Anatomical Taxonomy standardizes anatomical organization, and Ultrasonographic Diagnostic Attribute Framework formalizes nine diagnostic dimensions, including body system, organ, diagnosis, shape, margins, echogenicity, internal characteristics, posterior acoustic phenomena, and vascularity. Building upon these foundations, we propose Ultrasound-CLIP, a semantic-aware contrastive learning framework that introduces semantic soft labels and semantic loss to refine sample discrimination. Moreover, we construct a heterogeneous graph modality derived from UDAF's textual representations, enabling structured reasoning over lesion-attribute relations. Extensive experiments with patient-level data splitting demonstrate that our approach achieves state-of-the-art performance on classification and retrieval benchmarks, while also delivering strong generalization to zero-shot, linear probing, and fine-tuning tasks.