Sketch2CT: Multimodal Diffusion for Structure-Aware 3D Medical Volume Generation

arXiv cs.CV / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • Sketch2CT introduces a multimodal diffusion framework that generates structure-consistent 3D medical organ volumes by conditioning on a user-provided 2D sketch plus a textual description of 3D geometric semantics.
  • The method first produces anatomically consistent 3D segmentation masks from noise, using modules that refine sketch features with localized text cues and fuse global sketch–text representations via a capsule-attention backbone.
  • Generated segmentation masks are then used to guide a latent diffusion model for realistic 3D CT volume synthesis that matches the user-defined sketch and description.
  • Experiments on public CT datasets reportedly show improved performance over prior approaches, highlighting better multimodal controllability and reduced cost for medical dataset augmentation.
  • The project provides code publicly via GitHub, enabling researchers to test and build upon the proposed pipeline.

Abstract

Diffusion probabilistic models have demonstrated significant potential in generating high-quality, realistic medical images, providing a promising solution to the persistent challenge of data scarcity in the medical field. Nevertheless, producing 3D medical volumes with anatomically consistent structures under multimodal conditions remains a complex and unresolved problem. We introduce Sketch2CT, a multimodal diffusion framework for structure-aware 3D medical volume generation, jointly guided by a user-provided 2D sketch and a textual description that captures 3D geometric semantics. The framework initially generates 3D segmentation masks of the target organ from random noise, conditioned on both modalities. To effectively align and fuse these inputs, we propose two key modules that refine sketch features with localized textual cues and integrate global sketch-text representations. Built upon a capsule-attention backbone, these modules leverage the complementary strengths of sketches and text to produce anatomically accurate organ shapes. The synthesized segmentation masks subsequently guide a latent diffusion model for 3D CT volume synthesis, enabling realistic reconstruction of organ appearances that are consistent with user-defined sketches and descriptions. Extensive experiments on public CT datasets demonstrate that Sketch2CT achieves superior performance in generating multimodal medical volumes. Its controllable, low-cost generation pipeline enables principled, efficient augmentation of medical datasets. Code is available at https://github.com/adlsn/Sketch2CT.

Sketch2CT: Multimodal Diffusion for Structure-Aware 3D Medical Volume Generation | AI Navigate