Estimating 3D mesh of the human body from a single 2D image is an important task with many applications such as augmented reality and Human-Robot interaction. However, prior works reconstructed 3D mesh from global image feature extracted by using convolutional neural network (CNN), where the dense correspondences between the mesh surface and the image pixels are missing, leading to suboptimal solution. This paper proposes a model-free 3D human mesh estimation framework, named DecoMR, which explicitly establishes the dense correspondence between the mesh and the local image features in the UV space (i.e. a 2D space used for texture mapping of 3D mesh). DecoMR first predicts pixel-to-surface dense correspondence map (i.e., IUV image), with which we transfer local features from the image space to the UV space. Then the transferred local image features are processed in the UV space to regress a location map, which is well aligned with transferred features. Finally we reconstruct 3D human mesh from the regressed location map with a predefined mapping function. We also observe that the existing discontinuous UV map are unfriendly to the learning of network. Therefore, we propose a novel UV map that maintains most of the neighboring relations on the original mesh surface. Experiments demonstrate that our proposed local feature alignment and continuous UV map outperforms existing 3D mesh based methods on multiple public benchmarks. Code will be made available at https://github.com/zengwang430521/DecoMR
翻译:从单一 2D 图像中估算人体身体的 3D 网格是一个重要任务, 包括许多应用程序, 如扩大现实和人类- Robot 互动 。 然而, 先前的工程从全球图像特征中重建了 3D 网格, 利用 convolual 神经网络( CNN) 提取了3D 网格, 即网格表面和图像像素之间缺少密集的对应关系, 导致亚最佳解决方案 。 本文提出了一个名为 DecoMR 的无模型的 3D 人类网格估计框架, 它明确确定了紫外空间的网格和本地图像特征( 即用于3D Mesh 的纹理绘图的 2D 空间) 。 DecoMR首先预测了多层神经网络( 即 IUVV 图像), 我们将本地的本地功能从图像空间转移到拟议的位置图中, 然后将本地图像功能在紫外处理, 与已转移的地块图相匹配。 最后, 我们从未反向的地图地图上重建 3D 人类网格的地图,, 将维持我们现有的网络 。