Supervised Deep Learning (DL) models are currently the leading approach for sensor-based Human Activity Recognition (HAR) on wearable and mobile devices. However, training them requires large amounts of labeled data whose collection is often time-consuming, expensive, and error-prone. At the same time, due to the intra- and inter-variability of activity execution, activity models should be personalized for each user. In this work, we propose SelfAct: a novel framework for HAR combining self-supervised and active learning to mitigate these problems. SelfAct leverages a large pool of unlabeled data collected from many users to pre-train through self-supervision a DL model, with the goal of learning a meaningful and efficient latent representation of sensor data. The resulting pre-trained model can be locally used by new users, which will fine-tune it thanks to a novel unsupervised active learning strategy. Our experiments on two publicly available HAR datasets demonstrate that SelfAct achieves results that are close to or even better than the ones of fully supervised approaches with a small number of active learning queries.
翻译:监督式深度学习模型目前是可穿戴和移动设备上基于传感器的人体活动识别的主要方法。然而,这些模型的训练需要大量标记数据,数据的收集往往耗时、昂贵且容易出错。同时,由于活动执行的内部和间部变化,活动模型对每个用户应该进行个性化定制。在本文中,我们提出了SelfAct:一种将自我监督和主动学习结合起来从而缓解这些问题的新框架。SelfAct利用从许多用户收集的大量未标记数据通过自我监督预训练深度学习模型,以学习传感器数据的有意义、高效的潜在表示。由此产生的预训练模型可以被新用户本地使用,这些用户将通过一种新颖的无监督主动学习策略进行微调。我们在两个公开可用的人体活动识别数据集上的实验表明,SelfAct在少量主动学习查询下实现了与完全监督方法相当甚至更好的结果。