Robot-assisted surgery is an established clinical practice. The automatic identification of surgical actions is needed for a range of applications, including performance assessment of trainees and surgical process modeling for autonomous execution and monitoring. However, supervised action identification is not feasible, due to the burden of manually annotating recordings of potentially complex and long surgical executions. Moreover, often few example executions of a surgical procedure can be recorded. This paper proposes a novel algorithm for unsupervised identification of surgical actions in a standard surgical training task, the ring transfer, executed with da Vinci Research Kit. Exploiting kinematic and semantic visual features automatically extracted from a very limited dataset of executions, we are able to significantly outperform the state-of-the-art results for a similar application, improving the quality of segmentation (88% vs. 82% matching score) and clustering (67% vs. 54% F1-score) even in the presence of noise, short actions and non homogeneous workflows, i.e. non repetitive action sequences. Full action identification on hardware with standard commercial specifications is performed in less than 1 s for single execution.
翻译:机械辅助手术是一种既定的临床实践。 需要自动识别外科手术行动, 其中包括对受训人员进行性能评估, 以及用于自主执行和监测的外科手术程序模型。 但是, 监督行动识别不可行, 原因是人工注解潜在复杂和长期外科手术处决记录的负担。 此外, 通常很少能记录执行外科手术程序的例子。 本文提议在标准外科培训任务、 环状转移、 与da Vinci Research Kit 一起执行的外科手术手术手术手术手术手术操作的未经监督识别的新算法。 开发从非常有限的处决数据集中自动提取的动能和语义直观特征, 我们能够大大超过类似应用的最新结果, 提高分数质量( 88% 相对于 82% 匹配分数) 和聚群( 67% 对54% F1- 分数), 即使存在噪音、 短动作和不均匀的工作流程, 即不重复行动序列 。 对于符合标准商业规格的硬件的全面行动识别, 以不到 1 s s 。