The task of reconstructing 3D human motion has wideranging applications. The gold standard Motion capture (MoCap) systems are accurate but inaccessible to the general public due to their cost, hardware and space constraints. In contrast, monocular human mesh recovery (HMR) methods are much more accessible than MoCap as they take single-view videos as inputs. Replacing the multi-view Mo- Cap systems with a monocular HMR method would break the current barriers to collecting accurate 3D motion thus making exciting applications like motion analysis and motiondriven animation accessible to the general public. However, performance of existing HMR methods degrade when the video contains challenging and dynamic motion that is not in existing MoCap datasets used for training. This reduces its appeal as dynamic motion is frequently the target in 3D motion recovery in the aforementioned applications. Our study aims to bridge the gap between monocular HMR and multi-view MoCap systems by leveraging information shared across multiple video instances of the same action. We introduce the Neural Motion (NeMo) field. It is optimized to represent the underlying 3D motions across a set of videos of the same action. Empirically, we show that NeMo can recover 3D motion in sports using videos from the Penn Action dataset, where NeMo outperforms existing HMR methods in terms of 2D keypoint detection. To further validate NeMo using 3D metrics, we collected a small MoCap dataset mimicking actions in Penn Action,and show that NeMo achieves better 3D reconstruction compared to various baselines.
翻译:重建3D人类运动的任务涉及广泛的应用。金标准运动捕获系统(MoCap)准确,但由于成本、硬件和空间限制,公众无法使用。相比之下,单视人类网膜恢复方法比MoCap更容易获得,因为采用单视视频作为投入。用单视视频取代多视MCap系统,将打破目前收集准确3D运动的障碍,从而使公众能够使用运动分析和运动驱动动画等令人振奋的应用程序。然而,当视频包含挑战性和动态动作,而现有用于培训的MoCap数据集中没有这种动作时,现有HMR方法的性能会降低。相比之下,单视光人的网网网网网恢复方法要比MoCap多视网系统更容易获得。我们用多个视频实例共享的信息来弥补目前多视网的MCapta系统与多视网球运动之间的鸿沟。我们引入了Neural Momomo(NMM)字段的性能能代表3D运动背后的3D运动,我们用当前HMM 的MD 数据检测方法在目前运动中用HMR 3 的底图解图中,我们用HM 3 将现有数据校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校校的校的校的校的校里的数据进行了。