The matching of 3D shapes has been extensively studied for shapes represented as surface meshes, as well as for shapes represented as point clouds. While point clouds are a common representation of raw real-world 3D data (e.g. from laser scanners), meshes encode rich and expressive topological information, but their creation typically requires some form of (often manual) curation. In turn, methods that purely rely on point clouds are unable to meet the matching quality of mesh-based methods that utilise the additional topological structure. In this work we close this gap by introducing a self-supervised multimodal learning strategy that combines mesh-based functional map regularisation with a contrastive loss that couples mesh and point cloud data. Our shape matching approach allows to obtain intramodal correspondences for triangle meshes, complete point clouds, and partially observed point clouds, as well as correspondences across these data modalities. We demonstrate that our method achieves state-of-the-art results on several challenging benchmark datasets even in comparison to recent supervised methods, and that our method reaches previously unseen cross-dataset generalisation ability.
翻译:三维形状匹配已经在表面网格形状和点云形状中得到广泛研究。虽然点云是现实世界三维数据(例如来自激光扫描仪的数据)的常见表示方式,但网格编码了丰富和表达力强的拓扑信息,但其创建通常需要某种形式的(常常是手动的)管理。转而,纯粹依赖于点云的方法无法达到利用附加拓扑结构的基于网格的方法的匹配质量。在这项工作中,我们引入了一种自监督多模态学习策略,该策略结合了基于网格的函数映射正则化和一种对比损失,将网格和点云数据耦合在一起。我们的形状匹配方法允许获得针对三角形网格、完整点云和部分观测点云的单模态对应关系,以及跨这些数据模态的对应关系。我们在几个具有挑战性的基准数据集上展示了我们的方法达到了最先进的结果,甚至与最近的监督方法相比,并且我们的方法达到了以前未曾见过的跨数据集的泛化能力。