Since the application of Deep Q-Learning to the continuous action domain in Atari-like games, Deep Reinforcement Learning (Deep-RL) techniques for motion control have been qualitatively enhanced. Nowadays, modern Deep-RL can be successfully applied to solve a wide range of complex decision-making tasks for many types of vehicles. Based on this context, in this paper, we propose the use of Deep-RL to perform autonomous mapless navigation for Hybrid Unmanned Aerial Underwater Vehicles (HUAUVs), robots that can operate in both, air or water media. We developed two approaches, one deterministic and the other stochastic. Our system uses the relative localization of the vehicle and simple sparse range data to train the network. We compared our approaches with a traditional geometric tracking controller for mapless navigation. Based on experimental results, we can conclude that Deep-RL-based approaches can be successfully used to perform mapless navigation and obstacle avoidance for HUAUVs. Our vehicle accomplished the navigation in two scenarios, being capable to achieve the desired target through both environments, and even outperforming the geometric-based tracking controller on the obstacle-avoidance capability.
翻译:自深距离学习应用于Atari类游戏的持续行动领域以来,深度强化学习运动控制技术在质量上得到了提高。现在,现代深距离技术可以成功地用于解决许多类型车辆的一系列复杂决策任务。基于这一背景,我们提议使用深距离L为混合的无人驾驶空中水下潜水器(HUAUVs)进行自主无地图导航,这些飞行器可以在空气或水媒介中运作。我们开发了两种方法,一种是确定性方法,另一种是随机技术。我们的系统使用车辆相对定位和简单微小的射程数据来培训网络。我们将我们的方法与传统的无地图航行几何跟踪控制器进行比较。根据实验结果,我们可以得出结论,深距离导航仪可以成功地用于进行无地图的导航和避免HUAUVs的障碍。我们的飞行器在两种情景下完成了导航,能够通过环境实现预期的目标,甚至超越了以几何距离为主的轨道对障碍的跟踪能力。