Anatomical Prior-Driven Framework for Autonomous Robotic Cardiac Ultrasound Standard View Acquisition

arXiv cs.RO / 3/24/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an anatomical-prior (AP) driven framework to make autonomous robotic cardiac ultrasound standard view acquisition less dependent on operator skill.
  • It introduces a YOLO-based multi-class segmentation model enhanced with a Spatial-Relation Graph (SRG) module to embed anatomical priors into the feature pyramid for more consistent segmentation.
  • The method extracts quantifiable anatomical features from standard views, fits them to Gaussian distributions, and turns them into probabilistic APs used to guide scanning.
  • Robotic probe adjustment is reformulated as a reinforcement learning (RL) problem where the state is built from real-time anatomical features and the reward is based on how well the acquired anatomy matches the AP.
  • Experimental results report improved segmentation performance (SRG-YOLOv11s gains mAP50 by 11.3% and mIoU by 6.8%) and strong RL outcomes (92.5% success in simulation, 86.7% in phantom tests).

Abstract

Cardiac ultrasound diagnosis is critical for cardiovascular disease assessment, but acquiring standard views remains highly operator-dependent. Existing medical segmentation models often yield anatomically inconsistent results in images with poor textural differentiation between distinct feature classes, while autonomous probe adjustment methods either rely on simplistic heuristic rules or black-box learning. To address these issues, our study proposed an anatomical prior (AP)-driven framework integrating cardiac structure segmentation and autonomous probe adjustment for standard view acquisition. A YOLO-based multi-class segmentation model augmented by a spatial-relation graph (SRG) module is designed to embed AP into the feature pyramid. Quantifiable anatomical features of standard views are extracted. Their priors are fitted to Gaussian distributions to construct probabilistic APs. The probe adjustment process of robotic ultrasound scanning is formalized as a reinforcement learning (RL) problem, with the RL state built from real-time anatomical features and the reward reflecting the AP matching. Experiments validate the efficacy of the framework. The SRG-YOLOv11s improves mAP50 by 11.3% and mIoU by 6.8% on the Special Case dataset, while the RL agent achieves a 92.5% success rate in simulation and 86.7% in phantom experiments.