Surgical reconstruction of dynamic tissues from endoscopic videos is a crucial technology in robot-assisted surgery. The development of Neural Radiance Fields (NeRFs) has greatly advanced deformable tissue reconstruction, achieving high-quality results from video and image sequences. However, reconstructing deformable endoscopic scenes remains challenging due to aliasing and artifacts caused by tissue movement, which can significantly degrade visualization quality. The introduction of 3D Gaussian Splatting (3DGS) has improved reconstruction efficiency by enabling a faster rendering pipeline. Nevertheless, existing 3DGS methods often prioritize rendering speed while neglecting these critical issues. To address these challenges, we propose SAGS, a self-adaptive alias-free Gaussian splatting framework. We introduce an attention-driven, dynamically weighted 4D deformation decoder, leveraging 3D smoothing filters and 2D Mip filters to mitigate artifacts in deformable tissue reconstruction and better capture the fine details of tissue movement. Experimental results on two public benchmarks, EndoNeRF and SCARED, demonstrate that our method achieves superior performance in all metrics of PSNR, SSIM, and LPIPS compared to the state of the art while also delivering better visualization quality.
翻译:基于内窥镜视频的动态组织手术重建是机器人辅助手术中的关键技术。神经辐射场(NeRFs)的发展极大地推动了可变形组织重建的进展,能够从视频和图像序列中获得高质量结果。然而,由于组织运动引起的混叠和伪影,可变形内窥镜场景的重建仍然具有挑战性,这些因素会显著降低可视化质量。三维高斯溅射(3DGS)的引入通过实现更快的渲染管线,提高了重建效率。然而,现有的3DGS方法往往优先考虑渲染速度,而忽视了这些关键问题。为应对这些挑战,我们提出了SAGS,一种自适应无混叠高斯溅射框架。我们引入了一种注意力驱动的动态加权四维形变解码器,利用三维平滑滤波器和二维Mip滤波器来减轻可变形组织重建中的伪影,并更好地捕捉组织运动的精细细节。在EndoNeRF和SCARED两个公开基准测试上的实验结果表明,与现有最先进方法相比,我们的方法在PSNR、SSIM和LPIPS所有指标上均实现了更优的性能,同时提供了更好的可视化质量。