Ultrasound (US) imaging is commonly used to assist in the diagnosis and interventions of spine diseases, while the standardized US acquisitions performed by manually operating the probe require substantial experience and training of sonographers. In this work, we propose a novel dual-agent framework that integrates a reinforcement learning (RL) agent and a deep learning (DL) agent to jointly determine the movement of the US probe based on the real-time US images, in order to mimic the decision-making process of an expert sonographer to achieve autonomous standard view acquisitions in spinal sonography. Moreover, inspired by the nature of US propagation and the characteristics of the spinal anatomy, we introduce a view-specific acoustic shadow reward to utilize the shadow information to implicitly guide the navigation of the probe toward different standard views of the spine. Our method is validated in both quantitative and qualitative experiments in a simulation environment built with US data acquired from 17 volunteers. The average navigation accuracy toward different standard views achieves 5.18mm/5.25deg and 12.87mm/17.49deg in the intra- and inter-subject settings, respectively. The results demonstrate that our method can effectively interpret the US images and navigate the probe to acquire multiple standard views of the spine.
翻译:超声成像(US)通常用于协助诊断和干预脊椎疾病,而通过人工操作探头进行的美国标准化购置需要大量经验和对声学工作者的培训。在这项工作中,我们提议一个新型双重试剂框架,将强化学习剂和深学习剂结合起来,共同确定美国探头的移动情况,以实时美国图像为基础,以模拟一位专家的测音师的决策过程,从而实现脊椎声学的自主标准视图获取。此外,受美国传播性质和脊柱解剖特征的启发,我们引入了一种针对具体视景的声影奖赏,以利用影子信息暗中引导探头走向脊椎的不同标准观点。我们的方法在模拟环境中的定量和定性实验中得到了验证,以17名志愿者获得的数据为基础,不同标准视图的平均导航精确度分别为5.18mm/5.25deg和12.87毫米/17.49deg。此外,我们的方法可以有效地解释美国内部和跨主体环境中的图像和多面图象。