High-precision camera re-localization technology in a pre-established 3D environment map is the basis for many tasks, such as Augmented Reality, Robotics and Autonomous Driving. The point-based visual re-localization approaches are well-developed in recent decades, but are insufficient in some feature-less cases. In this paper, we propose a point-line joint optimization method for pose refinement with the help of the innovatively designed line extracting CNN named VLSE, and the line matching and pose optimization approach. We adopt a novel line representation and customize a hybrid convolutional block based on the Stacked Hourglass network, to detect accurate and stable line features on images. Then we apply a coarse-to-fine strategy to obtain precise 2D-3D line correspondences based on the geometric constraint. A following point-line joint cost function is constructed to optimize the camera pose with the initial coarse pose. Sufficient experiments are conducted on open datasets, i.e, line extractor on Wireframe and YorkUrban, localization performance on Aachen Day-Night v1.1 and InLoc, to confirm the effectiveness of our point-line joint pose optimization method.
翻译:在预先建立的三维环境地图中,高精度摄像头重新定位技术是许多任务的基础,例如增强现实、机器人和自主驱动等。近几十年来,基于点的视觉重新定位方法发展良好,但在某些无特色的案例中还不够。在本文中,我们提出一个点线联合优化方法,以便在创新设计的提取名为VLSE的CNN线以及线的匹配和优化方法的帮助下进行改进。我们采用了新颖的线代表制,并定制了一个以堆积玻璃网络为基础的混合共振区块,以探测图像上的准确和稳定的线性特征。然后,我们采用一个粗略到线的战略,以获得基于几何度限制的精确的 2D-3D 线对应。我们构建了一个以下的点线联合成本功能,以优化相机的配置,使其与最初的缩略图相相保持最佳化。在开放的数据集上进行了充分的实验,即Wirefram和YorkUrban的线提取器,在Aachen Day-Night v1.1和InLoc上进行地方化工作,以确认我们的联合优化方法的有效性。