Autonomous car racing is a challenging task in the robotic control area. Traditional modular methods require accurate mapping, localization and planning, which makes them computationally inefficient and sensitive to environmental changes. Recently, deep-learning-based end-to-end systems have shown promising results for autonomous driving/racing. However, they are commonly implemented by supervised imitation learning (IL), which suffers from the distribution mismatch problem, or by reinforcement learning (RL), which requires a huge amount of risky interaction data. In this work, we present a general deep imitative reinforcement learning approach (DIRL), which successfully achieves agile autonomous racing using visual inputs. The driving knowledge is acquired from both IL and model-based RL, where the agent can learn from human teachers as well as perform self-improvement by safely interacting with an offline world model. We validate our algorithm both in a high-fidelity driving simulation and on a real-world 1/20-scale RC-car with limited onboard computation. The evaluation results demonstrate that our method outperforms previous IL and RL methods in terms of sample efficiency and task performance. Demonstration videos are available at https://caipeide.github.io/autorace-dirl/
翻译:传统的模块化方法要求精确的绘图、本地化和规划,使得这些方法在计算上效率低,对环境变化敏感。最近,基于深学习的端到端系统显示了自主驾驶/赛跑的有希望的结果。然而,这些方法通常通过受监督的模拟学习(IL)加以实施,这种模拟存在分布错配问题,或者通过强化学习(RL)加以实施,这种学习需要大量风险互动数据。在这项工作中,我们介绍了一种一般的深效强化强化学习方法(DIRL),这种方法利用视觉投入成功地实现了快速自主赛。驱动知识来自IL和基于模型的RL,在那里,代理人可以向人类教师学习,并通过安全地与离线世界模型互动来进行自我改进。我们在高性驱动模拟中和在现实世界1/20规模的RC-car上验证我们的算法,而机上计算有限。评价结果表明,我们的方法在样本效率和任务性表现方面比以前IL和RL方法要好。演示录像可在 https://caipiderabir.imations.