Neural implicit representations are drawing a lot of attention from the robotics community recently, as they are expressive, continuous and compact. However, city-scale continual implicit dense mapping based on sparse LiDAR input is still an under-explored challenge. To this end, we successfully build a city-scale continual neural mapping system with a panoptic representation that consists of environment-level and instance-level modelling. Given a stream of sparse LiDAR point cloud, it maintains a dynamic generative model that maps 3D coordinates to signed distance field (SDF) values. To address the difficulty of representing geometric information at different levels in city-scale space, we propose a tailored three-layer sampling strategy to dynamically sample the global, local and near-surface domains. Meanwhile, to realize high fidelity mapping of instance under incomplete observation, category-specific prior is introduced to better model the geometric details. We evaluate on the public SemanticKITTI dataset and demonstrate the significance of the newly proposed three-layer sampling strategy and panoptic representation, using both quantitative and qualitative results. Codes and model will be publicly available.
翻译:神经隐式表示近来引起了机器人领域的广泛关注,因为它们是表达性强、连续和紧凑的。然而,在以稀疏LiDAR输入为基础的城市规模连续隐式稠密映射方面仍存在较大挑战。为此,我们成功构建了城市规模的连续神经映射系统,使用全景表示来实现环境级和实例级建模。给定一系列稀疏的LiDAR点云,系统维护了一个动态生成模型,将3D坐标映射为有符号距离场(SDF)值。为了解决在城市规模空间中表示不同层次几何信息的困难,我们提出了一种定制的三层采样策略,动态采样全球、局部和近表面域。同时,为了实现在不完整观测下实例的高保真映射,引入了类别特定的先验知识以更好地建模几何细节。我们在公开的SemanticKITTI数据集上进行评估,并使用定量和定性结果证明了新提出的三层采样策略和全景表示的重要性。代码和模型将公开提供。