Real-time scene reconstruction from depth data inevitably suffers from occlusion, thus leading to incomplete 3D models. Partial reconstructions, in turn, limit the performance of algorithms that leverage them for applications in the context of, e.g., augmented reality, robotic navigation, and 3D mapping. Most methods address this issue by predicting the missing geometry as an offline optimization, thus being incompatible with real-time applications. We propose a framework that ameliorates this issue by performing scene reconstruction and semantic scene completion jointly in an incremental and real-time manner, based on an input sequence of depth maps. Our framework relies on a novel neural architecture designed to process occupancy maps and leverages voxel states to accurately and efficiently fuse semantic completion with the 3D global model. We evaluate the proposed approach quantitatively and qualitatively, demonstrating that our method can obtain accurate 3D semantic scene completion in real-time.
翻译:从深度数据进行实时场景重建必然会受到封闭性的影响,从而导致3D模型的不完整。部分的重建反过来又限制了算法的性能,这种算法在扩大现实、机器人导航和3D绘图等方面运用这些算法来应用这些算法。大多数方法解决这一问题的方式是预测缺失的几何为离线优化,因此与实时应用不相容。我们提议了一个框架,通过在深度地图输入序列的基础上,以渐进和实时方式联合进行现场重建和语义场完成来缓解这一问题。我们的框架依靠一种新颖的神经结构,旨在处理占用图和杠杆 voxel 状态,以准确和高效地结合3D全球模型的语义完成。我们从数量和质量上评价了拟议的方法,表明我们的方法可以实时获得准确的 3D 语义场完成。