Self-supervised learning on point clouds has gained a lot of attention recently, since it addresses the label-efficiency and domain-gap problems on point cloud tasks. In this paper, we propose a novel self-supervised framework to learn informative representations from partial point clouds. We leverage partial point clouds scanned by LiDAR that contain both content and pose attributes, and we show that disentangling such two factors from partial point clouds enhances feature representation learning. To this end, our framework consists of three main parts: 1) a completion network to capture holistic semantics of point clouds; 2) a pose regression network to understand the viewing angle where partial data is scanned from; 3) a partial reconstruction network to encourage the model to learn content and pose features. To demonstrate the robustness of the learnt feature representations, we conduct several downstream tasks including classification, part segmentation, and registration, with comparisons against state-of-the-art methods. Our method not only outperforms existing self-supervised methods, but also shows a better generalizability across synthetic and real-world datasets.
翻译:点云自我监督的学习最近引起了许多关注,因为它解决了点云的标签效率和域网问题。 在本文中,我们提出一个新的自我监督框架,从局部云中学习信息陈述。我们利用LiDAR扫描的包含内容和构成属性的部分点云,我们表明,将这两个因素从局部云中分离出来会增加特征表述学习。为此,我们的框架由三个主要部分组成:1) 完成网络,以捕捉点云的整体语义;2) 构成回归网络,以了解部分数据从哪个角度扫描的视图;3) 部分重建网络,鼓励模型学习内容和设置特征。为显示所学习特征表述的稳健性,我们开展若干下游任务,包括分类、部分分割和注册,并与最新方法进行比较。我们的方法不仅超越了现有的自我监督方法,而且显示了合成和真实世界数据集的更普遍性。