Intraoperative shape reconstruction of organs from endoscopic camera images is a complex yet indispensable technique for image-guided surgery. To address the uncertainty in reconstructing entire shapes from single-viewpoint occluded images, we propose a framework for generative virtual learning of shape reconstruction using image translation with common latent variables between simulated and real images. As it is difficult to prepare sufficient amount of data to learn the relationship between endoscopic images and organ shapes, self-supervised virtual learning is performed using simulated images generated from statistical shape models. However, small differences between virtual and real images can degrade the estimation performance even if the simulated images are regarded as equivalent by humans. To address this issue, a Variational Autoencoder is used to convert real and simulated images into identical synthetic images. In this study, we targeted the shape reconstruction of collapsed lungs from thoracoscopic images and confirmed that virtual learning could improve the similarity between real and simulated images. Furthermore, shape reconstruction error could be improved by 16.9%.
翻译:从内窥镜相机图像对器官进行不合作的形状重建是一项复杂但不可或缺的图像制导外科手术技术。为了解决从单视点隐蔽图像中重建整个形状的不确定性,我们提出了一个框架,用于利用模拟图像与真实图像之间的共同潜伏变量转换图像,对形状重建进行基因化虚拟学习。由于难以编制足够的数据来了解内镜图像与器官形状之间的关系,因此使用从统计形状模型产生的模拟图像进行自我监督的虚拟学习。然而,虚拟图像与真实图像之间的微小差异可以降低估计性能,即使模拟图像被人类视为等效。为解决这一问题,将虚拟自动图解器用于将真实和模拟图像转换成相同的合成图像。在这项研究中,我们的目标是用硫镜图像来重建崩溃的肺的形状,并确认虚拟学习可以改善真实图像和模拟图像之间的相似性。此外,可将重建误差改善16.9 %。