Recent advances in machine learning showed that pre-training representations acquired via self-supervised learning could achieve high accuracy on tasks with small training data. Unlike in vision and natural language processing domains, such pre-training for IMU-based applications is challenging, as there are only a few publicly available datasets with sufficient size and diversity to learn generalizable representations. To overcome this problem, we propose IMG2IMU, a novel approach that adapts pre-train representation from large-scale images to diverse few-shot IMU sensing tasks. We convert the sensor data into visually interpretable spectrograms for the model to utilize the knowledge gained from vision. Further, we apply contrastive learning on an augmentation set we designed to learn representations that are tailored to interpreting sensor data. Our extensive evaluations on five different IMU sensing tasks show that IMG2IMU consistently outperforms the baselines, illustrating that vision knowledge can be incorporated into a few-shot learning environment for IMU sensing tasks.
翻译:最近在机器学习方面取得的进步表明,通过自我监督学习获得的培训前代表制在使用小型培训数据的任务上可以实现高度准确性。与视觉和自然语言处理领域不同,对IMU应用的这类培训前培训具有挑战性,因为只有少数一些公开可得的数据集,其尺寸和多样性都足以学习可概括性代表制。为解决这一问题,我们建议IMG2IMU采用一种新颖的方法,将大型图像的预培训代表制成多种微小的IMU遥感任务。我们把传感器数据转换成可视觉解读的光谱,用于模型,以利用从视觉中获取的知识。此外,我们用扩增版学习我们设计的用于解释传感器数据的显示制表。我们对IMG2IMU五个不同的IMU遥感任务的广泛评价表明,IMG2IMU的测测测任务始终超越了基线,表明愿景知识可以纳入IMU遥感任务的几发学环境。