Human action recognition is a well-known computer vision and pattern recognition task of identifying which action a man is actually doing. Extracting the keypoint information of a single human with both spatial and temporal features of action sequences plays an essential role to accomplish the task.In this paper, we propose a human action system for Red-Green-Blue(RGB) input video with our own designed module. Based on the efficient Gated Recurrent Unit(GRU) for spatio-temporal feature extraction, we add another sampling module and normalization module to improve the performance of the model in order to recognize the human actions. Furthermore, we build a novel dataset with a similar background and discriminative actions for both human keypoint prediction and behavior recognition. To get a better result, we retrain the pose model with our new dataset to get better performance. Experimental results demonstrate the effectiveness of the proposed model on our own human behavior recognition dataset and some public datasets.
翻译:人类行动识别是一项众所周知的计算机愿景和模式识别任务,即确定一个人实际正在从事何种行动。提取一个具有行动序列空间和时间特点的单一人类的关键点信息对于完成这项任务起着关键作用。 在本文件中,我们建议用我们自己设计模块的红色-绿色-蓝色(RGB)输入视频人的行动系统。根据高效的抽取时空特征的Gated 经常性单元(GRU),我们增加另一个抽样模块和正常化模块,以改善模型的性能,从而承认人类的行动。此外,我们为人类关键点预测和行为识别建立一个具有类似背景和歧视性行动的新型数据集。为了取得更好的结果,我们用我们的新数据集重新配置了假模型,以取得更好的性能。实验结果显示了我们人类行为识别数据集和一些公共数据集的拟议模型的有效性。