Localization is an essential task for mobile autonomous robotic systems that want to use pre-existing maps or create new ones in the context of SLAM. Today, many robotic platforms are equipped with high-accuracy 3D LiDAR sensors, which allow a geometric mapping, and cameras able to provide semantic cues of the environment. Segment-based mapping and localization have been applied with great success to 3D point-cloud data, while semantic understanding has been shown to improve localization performance in vision based systems. In this paper we combine both modalities in SemSegMap, extending SegMap into a segment based mapping framework able to also leverage color and semantic data from the environment to improve localization accuracy and robustness. In particular, we present new segmentation and descriptor extraction processes. The segmentation process benefits from additional distance information from color and semantic class consistency resulting in more repeatable segments and more overlap after re-visiting a place. For the descriptor, a tight fusion approach in a deep-learned descriptor extraction network is performed leading to a higher descriptiveness for landmark matching. We demonstrate the advantages of this fusion on multiple simulated and real-world datasets and compare its performance to various baselines. We show that we are able to find 50.9% more high-accuracy prior-less global localizations compared to SegMap on challenging datasets using very compact maps while also providing accurate full 6 DoF pose estimates in real-time.
翻译:本地化是移动自主机器人系统的一项基本任务,这些系统希望使用已有的地图或在SLAM M范围内创建新的地图。 今天,许多机器人平台都配备了高精度的 3D LiDAR 传感器,可以进行几何映射,相机能够提供环境的语义提示。基于部分的绘图和本地化应用为3D 点球数据带来了极大的成功,而语义理解则表明可以改善基于愿景的系统中的本地化性能。在本文中,我们将SemSeSegMap中两种方法结合起来,将SegMap扩展成一个基于部分的绘图框架,能够同时利用环境的颜色和语义数据来提高本地化的准确性和稳健性。特别是,我们介绍了新的分解和脱义提取过程。基于颜色和语义类一致性的额外远程信息对3D点数据产生了益处,从而在重新访问一个地方后又出现了更多的重复部分和重叠。对于淡化的描述,在深层次的描述网络中,一种紧密的聚合方法正在进行,从而提高基于环境的颜色和语义性数据精确度的描述性框架,同时我们可以将其展示其高性地标性地标比前的优势。我们展示了该级数据展示了多级数据。