Depth completion deals with the problem of recovering dense depth maps from sparse ones, where color images are often used to facilitate this completion. Recent approaches mainly focus on image guided learning to predict dense results. However, blurry image guidance and object structures in depth still impede the performance of image guided frameworks. To tackle these problems, we explore a repetitive design in our image guided network to sufficiently and gradually recover depth values. Specifically, the repetition is embodied in a color image guidance branch and a depth generation branch. In the former branch, we design a repetitive hourglass network to extract higher-level image features of complex environments, which can provide powerful context guidance for depth prediction. In the latter branch, we design a repetitive guidance module based on dynamic convolution where the convolution factorization is applied to simultaneously reduce its complexity and progressively model high-frequency structures, e.g., boundaries. Further, in this module, we propose an adaptive fusion mechanism to effectively aggregate multi-step depth features. Extensive experiments show that our method achieves state-of-the-art result on the NYUv2 dataset and ranks 1st on the KITTI benchmark at the time of submission.
翻译:深度完成涉及从稀薄的深度地图中恢复密集的深度地图的问题,这些地图往往使用彩色图像来帮助完成。最近的方法主要侧重于图像引导学习,以预测密集的结果。然而,模糊的图像指导以及深度对象结构仍然阻碍着图像引导框架的性能。为了解决这些问题,我们探索了我们的图像引导网络中的重复设计,以足够和逐步恢复深度值。具体地说,重复体现在一个彩色图像指导分支和一个深度生成分支中。在前分支中,我们设计了一个重复的沙漏网络,以提取复杂环境的更高层次图像特征,为深度预测提供强有力的背景指导。在后一个分支中,我们设计了一个重复性的指导模块,以动态共进化为基础,同时使用共进系数来减少其复杂性并逐步建模高频结构,例如边界。此外,在这个模块中,我们提出了一个适应性融合机制,以有效地整合多步骤深度特征。广泛的实验表明,我们的方法在提交时达到了NYUv2数据集的状态,并在KITTI基准排名第1位。