Comprehending human motion is a fundamental challenge for developing Human-Robot Collaborative applications. Computer vision researchers have addressed this field by only focusing on reducing error in predictions, but not taking into account the requirements to facilitate its implementation in robots. In this paper, we propose a new model based on Transformer that simultaneously deals with the real time 3D human motion forecasting in the short and long term. Our 2-Channel Transformer (2CH-TR) is able to efficiently exploit the spatio-temporal information of a shortly observed sequence (400ms) and generates a competitive accuracy against the current state-of-the-art. 2CH-TR stands out for the efficient performance of the Transformer, being lighter and faster than its competitors. In addition, our model is tested in conditions where the human motion is severely occluded, demonstrating its robustness in reconstructing and predicting 3D human motion in a highly noisy environment. Our experiment results show that the proposed 2CH-TR outperforms the ST-Transformer, which is another state-of-the-art model based on the Transformer, in terms of reconstruction and prediction under the same conditions of input prefix. Our model reduces in 8.89% the mean squared error of ST-Transformer in short-term prediction, and 2.57% in long-term prediction in Human3.6M dataset with 400ms input prefix. Visit our website $\href{https://sites.google.com/view/estevevallsmascaro/publications/iros2022}{here}$.
翻译:理解人体动作是开发人机协作应用的基本挑战。计算机视觉研究人员通过仅关注预测误差来解决这个领域,但并没有考虑到其在机器人应用中实施所需的要求。在本文中,我们提出了一种基于Transformer的新模型,能够同时处理短期和长期下的实时三维人体动作预测。我们的二通道Transformer (2CH-TR) 能够有效地利用短时间观察序列(400ms)的时空信息,并在当前最先进的方法之间生成具有竞争力的准确性。2CH-TR凭借着Transformer的高效性能脱颖而出,比竞争对手更加轻量化、更快速。此外,我们的模型经过了在人体动作严重遮挡的情况下的测试,证明了其在高噪声环境中重构和预测3D人体动作的稳健性。我们的实验结果表明,所提出的2CH-TR在相同输入前缀条件下,在重建和预测方面优于基于Transformer的另一个当前最先进的模型ST-Transformer。在400ms输入前缀的Human3.6M数据集中,我们的模型将ST-Transformer在短期预测中的均方误差降低了8.89%,在长期预测中降低了2.57%。请访问我们的网站$\href{https://sites.google.com/view/estevevallsmascaro/publications/iros2022}{这里}$。