Drones have become a common tool, which is utilized in many tasks such as aerial photography, surveillance, and delivery. However, operating a drone requires more and more interaction with the user. A natural and safe method for Human-Drone Interaction (HDI) is using gestures. In this paper, we introduce an HDI framework building upon skeleton-based pose estimation. Our framework provides the functionality to control the movement of the drone with simple arm gestures and to follow the user while keeping a safe distance. We also propose a monocular distance estimation method, which is entirely based on image features and does not require any additional depth sensors. To perform comprehensive experiments and quantitative analysis, we create a customized testing dataset. The experiments indicate that our HDI framework can achieve an average of 93.5\% accuracy in the recognition of 11 common gestures. The code is available at: https://github.com/Zrrr1997/Pose2Drone
翻译:无人驾驶飞机操作需要与用户进行越来越多的互动。人类-日光互动的自然和安全方法正在使用手势。本文介绍一个基于骨架的人类人类发展倡议框架,以基于骨架的表面估计值为基础。我们的框架提供了以简单的手臂手势控制无人驾驶飞机移动并跟踪用户的功能,同时保持安全距离。我们还提出了一个单目距离估计方法,它完全基于图像特征,不需要额外的深度传感器。为了进行全面试验和定量分析,我们创建了一个定制的测试数据集。实验表明,我们的人类发展倡议框架在承认11个常见手势的情况下,可以达到平均93.5 ⁇ 的准确度。代码见:https://github.com/Zrr1997/Pose2Drone。