Tissue tracking plays a critical role in various surgical navigation and extended reality (XR) applications. While current methods trained on large synthetic datasets achieve high tracking accuracy and generalize well to endoscopic scenes, their runtime performances fail to meet the low-latency requirements necessary for real-time surgical applications. To address this limitation, we propose LiteTracker, a low-latency method for tissue tracking in endoscopic video streams. LiteTracker builds on a state-of-the-art long-term point tracking method, and introduces a set of training-free runtime optimizations. These optimizations enable online, frame-by-frame tracking by leveraging a temporal memory buffer for efficient feature reuse and utilizing prior motion for accurate track initialization. LiteTracker demonstrates significant runtime improvements being around 7x faster than its predecessor and 2x than the state-of-the-art. Beyond its primary focus on efficiency, LiteTracker delivers high-accuracy tracking and occlusion prediction, performing competitively on both the STIR and SuPer datasets. We believe LiteTracker is an important step toward low-latency tissue tracking for real-time surgical applications in the operating room. Our code is publicly available at https://github.com/ImFusionGmbH/lite-tracker.
翻译:组织追踪在各种手术导航与扩展现实(XR)应用中扮演着关键角色。尽管当前基于大规模合成数据集训练的方法在追踪精度上表现优异,并能良好泛化至内窥镜场景,但其运行时性能无法满足实时手术应用所需的低延迟要求。为应对这一局限,本文提出LiteTracker,一种面向内窥镜视频流的低延迟组织追踪方法。LiteTracker基于先进的长期点追踪方法构建,并引入一系列无需训练的运行优化策略。这些优化通过利用时序记忆缓冲区实现高效特征复用,并借助先验运动信息进行精确轨迹初始化,从而支持在线逐帧追踪。实验表明,LiteTracker在运行效率上取得显著提升:相比其前代方法提速约7倍,较当前最优方法快2倍。除注重效率外,LiteTracker同时实现了高精度追踪与遮挡预测,在STIR和SuPer数据集上均展现出竞争力。我们相信LiteTracker是迈向手术室实时低延迟组织追踪的重要一步。代码已公开于https://github.com/ImFusionGmbH/lite-tracker。