Obstacle avoidance is a fundamental and challenging problem for autonomous navigation of mobile robots. In this paper, we consider the problem of obstacle avoidance in simple 3D environments where the robot has to solely rely on a single monocular camera. In particular, we are interested in solving this problem without relying on localization, mapping, or planning techniques. Most of the existing work consider obstacle avoidance as two separate problems, namely obstacle detection, and control. Inspired by the recent advantages of deep reinforcement learning in Atari games and understanding highly complex situations in Go, we tackle the obstacle avoidance problem as a data-driven end-to-end deep learning approach. Our approach takes raw images as input and generates control commands as output. We show that discrete action spaces are outperforming continuous control commands in terms of expected average reward in maze-like environments. Furthermore, we show how to accelerate the learning and increase the robustness of the policy by incorporating predicted depth maps by a generative adversarial network.
翻译:避免障碍是移动机器人自主导航的一个根本性和具有挑战性的问题。 在本文中,我们考虑了在简单3D环境中避免障碍的问题,在这种环境中,机器人必须完全依靠单一的单色相机。特别是,我们有兴趣解决这一问题,而不必依赖本地化、绘图或规划技术。大多数现有工作认为避免障碍是两个不同的问题,即发现障碍和控制。在阿塔里游戏中深入强化学习和理解戈内高度复杂的情况的最近优势的启发下,我们把避免障碍问题作为数据驱动的端到端深学习方法来处理。我们的方法将原始图像作为输入,并生成控制命令作为输出。我们表明,在像迷宫一样的环境中,离散行动空间在预期的平均报酬方面比连续控制命令要强。此外,我们通过一个基因化对抗网络,展示如何加速学习和提高政策的稳健性,将预测深度地图纳入到一起。