Many problems in machine learning involve regressing outputs that do not lie on a Euclidean space, such as a discrete probability distribution, or the pose of an object. An approach to tackle these problems through gradient-based learning consists in including in the deep learning architecture a differentiable function mapping arbitrary inputs of a Euclidean space onto this manifold. In this work, we establish a set of properties that such mapping should satisfy to allow proper training, and illustrate it in the case of 3D rotations. Through theoretical considerations and methodological experiments on a variety of tasks, we compare various differentiable mappings on the 3D rotation space, and conjecture about the importance of the local linearity of the mapping. We notably show that a mapping based on Procrustes orthonormalization of a 3x3 matrix generally performs best among the ones considered, but that rotation-vector representation might also be suitable when restricted to small angles.
翻译:机器学习的许多问题涉及不存在于欧几里德空间的递减产出,例如离散概率分布或物体的外形。通过梯度学习解决这些问题的方法包括:在深层学习结构中包括一种可区分的功能,将欧几里德空间的任意输入绘图到这一方块上。在这项工作中,我们建立了一套这种绘图应该满足的属性,以便能够进行适当的培训,并在3D轮作中加以说明。通过对各种任务进行理论考虑和方法实验,我们比较了3D旋转空间上的各种不同绘图,并推测了绘制图的本地直线性的重要性。我们特别表明,基于3x3矩阵的普罗克鲁斯特斯或正统化的绘图一般在所考虑的方位中表现最佳,但是,在限于小角度时,轮换-视频代表也可能是合适的。