Visual localization is a fundamental task that regresses the 6 Degree Of Freedom (6DoF) poses with image features in order to serve the high precision localization requests in many robotics applications. Degenerate conditions like motion blur, illumination changes and environment variations place great challenges in this task. Fusion with additional information, such as sequential information and Inertial Measurement Unit (IMU) inputs, would greatly assist such problems. In this paper, we present an efficient client-server visual localization architecture that fuses global and local pose estimations to realize promising precision and efficiency. We include additional geometry hints in mapping and global pose regressing modules to improve the measurement quality. A loosely coupled fusion policy is adopted to leverage the computation complexity and accuracy. We conduct the evaluations on two typical open-source benchmarks, 4Seasons and OpenLORIS. Quantitative results prove that our framework has competitive performance with respect to other state-of-the-art visual localization solutions.
翻译:视觉本地化是一项基本任务,它使6度自由(6DoF)的图像特性倒退,以便在许多机器人应用中满足高度精确的本地化要求。运动模糊、照明变化和环境变异等退化条件给这项任务带来了巨大的挑战。如果加上更多的信息,如相继信息和惰性测量股(IMU)的投入,将大大有助于这些问题。在本文件中,我们提出了一个高效的客户-服务器视觉本地化结构,将全球和本地的估算结合起来,以实现有希望的精确性和效率。我们在绘图中增加了几何学提示,而全球的图像回归模块则提高测量质量。我们采用了一种松散的混合政策来利用计算的复杂性和准确性。我们根据两个典型的公开源基准,即4Searsons和OpenLORIS进行评估。定量结果证明,我们的框架在其它状态的视觉本地化解决方案上具有竞争性的绩效。