Gesture recognition is one of the most intuitive ways of interaction and has gathered particular attention for human computer interaction. Radar sensors possess multiple intrinsic properties, such as their ability to work in low illumination, harsh weather conditions, and being low-cost and compact, making them highly preferable for a gesture recognition solution. However, most literature work focuses on solutions with a limited range that is lower than a meter. We propose a novel architecture for a long-range (1m - 2m) gesture recognition solution that leverages a point cloud-based cross-learning approach from camera point cloud to 60-GHz FMCW radar point cloud, which allows learning better representations while suppressing noise. We use a variant of Dynamic Graph CNN (DGCNN) for the cross-learning, enabling us to model relationships between the points at a local and global level and to model the temporal dynamics a Bi-LSTM network is employed. In the experimental results section, we demonstrate our model's overall accuracy of 98.4% for five gestures and its generalization capability.
翻译:雷达传感器具有多种内在特性,例如能够在低照明度、恶劣的天气条件下工作,而且成本低、紧凑,因此非常适合姿态识别解决方案。然而,大多数文献工作的重点是范围小于一米的有限解决方案。我们提出了远程(1米-2米)姿态识别新结构,利用基于云的点对点交叉学习方法,从摄像点云到60千兆赫调频CW雷达点云,在抑制噪音的同时学习更好的表现。我们用动态CNN(DGCNN)变量进行交叉学习,从而能够模拟地方和全球层面各点之间的关系,并模拟使用Bi-LSTM网络的时间动态。在实验结果一节中,我们展示了我们模型的98.4%的总体精确度,用于五个姿态及其一般化能力。