It is known that the kinematics of the human body skeleton reveals valuable information in action recognition. Recently, modeling skeletons as spatio-temporal graphs with Graph Convolutional Networks (GCNs) has been reported to solidly advance the state-of-the-art performance. However, GCN-based approaches exclusively learn from raw skeleton data, and are expected to extract the inherent structural information on their own. This paper describes REGINA, introducing a novel way to REasoning Graph convolutional networks IN Human Action recognition. The rationale is to provide to the GCNs additional knowledge about the skeleton data, obtained by handcrafted features, in order to facilitate the learning process, while guaranteeing that it remains fully trainable in an end-to-end manner. The challenge is to capture complementary information over the dynamics between consecutive frames, which is the key information extracted by state-of-the-art GCN techniques. Moreover, the proposed strategy can be easily integrated in the existing GCN-based methods, which we also regard positively. Our experiments were carried out in well known action recognition datasets and enabled to conclude that REGINA contributes for solid improvements in performance when incorporated to other GCN-based approaches, without any other adjustment regarding the original method. For reproducibility, the REGINA code and all the experiments carried out will be publicly available at https://github.com/DegardinBruno.
翻译:众所周知,人体骨骼的传动特征揭示了行动认知中的宝贵信息。最近,以图表革命网络(GCNs)为基本数据模型的模型骨架与时速图相比,以图表革命网络(GCNs)为模型骨架为模型,以稳步推进最新业绩;然而,基于GCN的方法仅从原始骨架数据中学习,预计将自行提取固有的结构信息。本文描述了REGINA, 引入了在人类行动认知中重新发现图表共振网络的新方式。其理由是向GCN提供以手工制作的特征获得的骨架数据的额外知识,以促进学习进程,同时保证它仍然能够以端到端的方式充分培训。但挑战在于掌握连续框架之间的动态的补充信息,即由最新GCN技术获得的关键信息。此外,拟议的战略可以很容易地融入现有的GCN-CGNG-Celgs 方法,我们也对此予以肯定。我们的实验是在众所周知的行动识别数据集中进行更多的行动识别数据,并且能够在RGAGINA的所有原始方法中进行真正的改进。