Vision-based bronchoscopy (VB) models require the registration of the virtual lung model with the frames from the video bronchoscopy to provide effective guidance during the biopsy. The registration can be achieved by either tracking the position and orientation of the bronchoscopy camera or by calibrating its deviation from the pose (position and orientation) simulated in the virtual lung model. Recent advances in neural networks and temporal image processing have provided new opportunities for guided bronchoscopy. However, such progress has been hindered by the lack of comparative experimental conditions. In the present paper, we share a novel synthetic dataset allowing for a fair comparison of methods. Moreover, this paper investigates several neural network architectures for the learning of temporal information at different levels of subject personalization. In order to improve orientation measurement, we also present a standardized comparison framework and a novel metric for camera orientation learning. Results on the dataset show that the proposed metric and architectures, as well as the standardized conditions, provide notable improvements to current state-of-the-art camera pose estimation in video bronchoscopy.
翻译:以视觉为基础的支气管镜(VB)模型要求将虚拟肺型模型与视频支气管镜(VB)的框架进行注册,以便在生物检查期间提供有效的指导。通过跟踪支气管照相机的位置和方向或校准其偏离虚拟肺型模型模拟的外形(位置和方向),可以实现注册。神经网络和时间图像处理方面的最新进展为指导支气管镜提供了新的机会。然而,由于缺乏比较实验条件,这种进展受到阻碍。在本文件中,我们分享了一套新的合成数据集,以便能够对方法进行公平的比较。此外,本文还调查了在不同层次的主题个性化中学习时间信息的若干神经网络结构。为了改进定向测量,我们还提出了一个标准化比较框架和新的相机定向学习指标。关于数据集的结果显示,拟议的指标和结构以及标准化条件为目前最先进的照相机在视频支气管镜检查中作出估计提供了显著的改进。