Knowledge transfer from synthetic to real data has been widely studied to mitigate data annotation constraints in various computer vision tasks such as semantic segmentation. However, the study focused on 2D images and its counterpart in 3D point clouds segmentation lags far behind due to the lack of large-scale synthetic datasets and effective transfer methods. We address this issue by collecting SynLiDAR, a large-scale synthetic LiDAR dataset that contains point-wise annotated point clouds with accurate geometric shapes and comprehensive semantic classes. SynLiDAR was collected from multiple virtual environments with rich scenes and layouts which consists of over 19 billion points of 32 semantic classes. In addition, we design PCT, a novel point cloud translator that effectively mitigates the gap between synthetic and real point clouds. Specifically, we decompose the synthetic-to-real gap into an appearance component and a sparsity component and handle them separately which improves the point cloud translation greatly. We conducted extensive experiments over three transfer learning setups including data augmentation, semi-supervised domain adaptation and unsupervised domain adaptation. Extensive experiments show that SynLiDAR provides a high-quality data source for studying 3D transfer and the proposed PCT achieves superior point cloud translation consistently across the three setups. SynLiDAR project page: \url{https://github.com/xiaoaoran/SynLiDAR}
翻译:合成数据向真实数据转移知识的做法已经得到广泛研究,以减轻诸如语义部分等各种计算机视觉任务的数据说明限制,然而,研究侧重于2D图像,3D点云分解的对应方则由于缺少大规模合成数据集和有效的传输方法而远远落后于2D图像,而3D点云分解则缺乏大规模合成数据集和有效的传输方法。我们通过收集大型合成激光雷达数据集(SynLiDAR)来解决这一问题,该数据集包含有精确的几何形状和全面语义分类的点注解云。SyLiDAR是从多个虚拟环境中收集的,该环境中有丰富的场景和布局,由超过190亿点的32个语义类组成。此外,我们设计了PCT,这是一个新的点云转换器,能够有效地缩小合成和真实点云云云云层之间的鸿沟。我们通过收集合成到一个外观组成部分和一个宽度部分,并分别处理这些部分,从而大大改进点云层翻译。我们对三个传输学习组进行了广泛的实验,包括数据增强、半监督域适应和不受监控域域适应。我们设计了190亿分的域适应。我们设计设计PLD。广域实验显示SyLD在SyLD上持续进行高质量的翻译。