While much progress has been made on the task of 3D point cloud registration, there still exists no learning-based method able to estimate the 6D pose of an object observed by a 2.5D sensor in a scene. The challenges of this scenario include the fact that most measurements are outliers depicting the object's surrounding context, and the mismatch between the complete 3D object model and its self-occluded observations. We introduce the first deep learning framework capable of effectively handling this scenario. Our method consists of an instance segmentation module followed by a pose estimation one. It allows us to perform 3D registration in a one-shot manner, without requiring an expensive iterative procedure. We further develop an on-the-fly rendering-based training strategy that is both time- and memory-efficient. Our experiments evidence the superiority of our approach over the state-of-the-art traditional and learning-based 3D registration methods.
翻译:虽然在3D点云登记任务上取得了很大进展,但是仍然没有基于学习的方法能够估计2.5D传感器在现场观察到的物体的6D构成,这种情景的挑战包括:大多数测量结果都是描述物体周围环境的外星,以及完整的3D物体模型与其自我封闭的观测结果不匹配。我们引入了第一个能够有效处理这一情景的深层次学习框架。我们的方法包括一个实例分割模块,然后是一种构成估计。它使我们能够以一拍方式进行3D登记,而不需要昂贵的迭接程序。我们进一步制定了一个基于飞行提供的培训战略,既具有时间效率,又具有记忆效率。我们的实验证明我们的方法优于最先进的传统和基于学习的3D登记方法。