This paper introduces a novel multi-view 6 DoF object pose refinement approach focusing on improving methods trained on synthetic data. It is based on the DPOD detector, which produces dense 2D-3D correspondences between the model vertices and the image pixels in each frame. We have opted for the use of multiple frames with known relative camera transformations, as it allows introduction of geometrical constraints via an interpretable ICP-like loss function. The loss function is implemented with a differentiable renderer and is optimized iteratively. We also demonstrate that a full detection and refinement pipeline, which is trained solely on synthetic data, can be used for auto-labeling real data. We perform quantitative evaluation on LineMOD, Occlusion, Homebrewed and YCB-V datasets and report excellent performance in comparison to the state-of-the-art methods trained on the synthetic and real data. We demonstrate empirically that our approach requires only a few frames and is robust to close camera locations and noise in extrinsic camera calibration, making its practical usage easier and more ubiquitous.
翻译:本文介绍了一个新的多视图 6 DoF 对象的改进方法,侧重于改进合成数据培训的方法,它以DPOD探测器为基础,该探测器在模型顶端和每个框架的图像像素之间产生密集的 2D-3D 对应。我们选择使用已知相对摄像机变形的多框架,因为它允许通过可解释的比较方案类似的损失函数引入几何限制。损失功能是用一个不同的翻版器执行的,并且是迭接式的优化。我们还表明,完全探测和精炼管道(仅对合成数据进行培训)可用于自动标注真实数据。我们对LineMOD、Oclusion、Homebred和YCB-V数据集进行定量评估,并报告与合成和真实数据培训的最新方法相比的优异性。我们从经验上表明,我们的方法只需要几个框架,并且能够紧紧地关闭外部摄像器的摄像点和噪音,从而使其实际使用更加容易和清晰。