Image guided depth completion aims to recover per-pixel dense depth maps from sparse depth measurements with the help of aligned color images, which has a wide range of applications from robotics to autonomous driving. However, the 3D nature of sparse-to-dense depth completion has not been fully explored by previous methods. In this work, we propose a Graph Convolution based Spatial Propagation Network (GraphCSPN) as a general approach for depth completion. First, unlike previous methods, we leverage convolution neural networks as well as graph neural networks in a complementary way for geometric representation learning. In addition, the proposed networks explicitly incorporate learnable geometric constraints to regularize the propagation process performed in three-dimensional space rather than in two-dimensional plane. Furthermore, we construct the graph utilizing sequences of feature patches, and update it dynamically with an edge attention module during propagation, so as to better capture both the local neighboring features and global relationships over long distance. Extensive experiments on both indoor NYU-Depth-v2 and outdoor KITTI datasets demonstrate that our method achieves the state-of-the-art performance, especially when compared in the case of using only a few propagation steps. Code and models are available at the project page.
翻译:图像引导深度完成的目的是,在对齐色图像的帮助下,从稀薄的深深度测量中恢复每像素密密密密深的深度地图,这种测量具有从机器人到自主驱动的广泛应用。然而,以前的方法并未充分探索稀到感知深度完成的三维性质。在这项工作中,我们提议以基于图表的集成空间推进网络(GraphCSPN)作为深度完成的一般方法。首先,与以往的方法不同,我们利用卷动神经网络和图形神经网络,以互补的方式进行几何代表性学习。此外,拟议的网络明确纳入可学习的几何限制,以规范在三维空间而不是二维平面上进行的传播过程。此外,我们利用地貌补丁序列来构建图,并在传播过程中用边距模块进行动态更新,以便更好地捕捉本地相邻特征和长距离的全球关系。在室内NYU-Deph-v2和户外KITTI数据集上进行的广泛实验,表明我们的方法只能使用数种状态模型,特别是在比较软件的版本中,只有几个步骤。