Human Activity Recognition (HAR) using on-body devices identifies specific human actions in unconstrained environments. HAR is challenging due to the inter and intra-variance of human movements; moreover, annotated datasets from on-body devices are scarce. This problem is mainly due to the difficulty of data creation, i.e., recording, expensive annotation, and lack of standard definitions of human activities. Previous works demonstrated that transfer learning is a good strategy for addressing scenarios with scarce data. However, the scarcity of annotated on-body device datasets remains. This paper proposes using datasets intended for human-pose estimation as a source for transfer learning; specifically, it deploys sequences of annotated pixel coordinates of human joints from video datasets for HAR and human pose estimation. We pre-train a deep architecture on four benchmark video-based source datasets. Finally, an evaluation is carried out on three on-body device datasets improving HAR performance.
翻译:使用人体活动识别(HAR)使用机体装置确定在不受限制的环境中的具体人类行动。HAR具有挑战性,因为人类运动之间和内部的变化;此外,机体装置的附加说明的数据集很少。这个问题主要是由于数据创建的困难,即记录、昂贵的批注和缺乏人类活动的标准定义。以前的工作表明,转让学习是用稀缺数据处理假设情景的良好战略。然而,附带说明的机体装置数据集仍然稀缺。本文提议使用用于人为用途估计的数据集作为转移学习的来源;具体地说,它利用HAR和人体表面估计视频数据集的附加说明的人类联合像素坐标序列。我们预先设计了四个基于视频的基准源数据集的深层结构。最后,对改进HAR性能的三个机体装置数据集进行了评价。