Professional race car drivers can execute extreme overtaking maneuvers. However, conventional systems for autonomous overtaking rely on either simplified assumptions about the vehicle dynamics or solving expensive trajectory optimization problems online. When the vehicle is approaching its physical limits, existing model-based controllers struggled to handle highly nonlinear dynamics and cannot leverage the large volume of data generated by simulation or real-world driving. To circumvent these limitations, this work proposes a new learning-based method to tackle the autonomous overtaking problem. We evaluate our approach using Gran Turismo Sport -- a world-leading car racing simulator known for its detailed dynamic modeling of various cars and tracks. By leveraging curriculum learning, our approach leads to faster convergence as well as increased performance compared to vanilla reinforcement learning. As a result, the trained controller outperforms the built-in model-based game AI and achieves comparable overtaking performance with an experienced human driver.
翻译:职业赛车驾驶员可以实施极端超载操作。然而,传统的自动超载系统依赖于对车辆动态的简化假设或解决在线上昂贵的轨迹优化问题。当车辆接近物理极限时,现有基于模型的控制员努力处理高度非线性动态,无法利用模拟或现实世界驱动产生的大量数据。为绕过这些限制,这项工作提出了一种新的基于学习的方法来解决自动超载问题。我们用Gran Turismo Sport -- -- 一个以其各种汽车和轨道的详细动态模型而闻名的世界领先的赛车模拟器 -- -- 来评估我们的方法。通过利用课程学习,我们的方法可以更快地趋同并增加与香草强化学习相比的性能。结果,受过训练的控制员超越了基于模型的内置游戏AI,实现了与有经验的人驾驶员可比的超载性性能。