This paper describes how advanced deep learning based computer vision algorithms are applied to enable real-time on-board sensor processing for small UAVs. Four use cases are considered: target detection, classification and localization, road segmentation for autonomous navigation in GNSS-denied zones, human body segmentation, and human action recognition. All algorithms have been developed using state-of-the-art image processing methods based on deep neural networks. Acquisition campaigns have been carried out to collect custom datasets reflecting typical operational scenarios, where the peculiar point of view of a multi-rotor UAV is replicated. Algorithms architectures and trained models performances are reported, showing high levels of both accuracy and inference speed. Output examples and on-field videos are presented, demonstrating models operation when deployed on a GPU-powered commercial embedded device (NVIDIA Jetson Xavier) mounted on board of a custom quad-rotor, paving the way to enabling high level autonomy.
翻译:本文介绍了如何应用先进的深层学习计算机视觉算法,以便对小型无人驾驶航空器进行实时机载传感器处理。考虑了四种使用案例:目标探测、分类和本地化、全球导航卫星系统封闭区的自主导航路段分割、人体身体分解和人类行动识别。所有算法都是利用以深神经网络为基础的最先进的图像处理方法开发的。已经开展了采购活动,以收集反映典型操作情景的定制数据集,其中复制了多旋式无人驾驶航空器的独特观点。报告了阿尔戈里特姆斯建筑和经过培训的模型性能,显示了高度的精确度和推断速度。提供了输出实例和现场视频,演示了安装在自定义的夸德罗托机上安装的GUP动力商业嵌入装置(NVIDIA Jetson Xavier)上的模型操作模式,从而为高水平的自主性铺平了道路。