When a vehicle drives on the road, its behaviors will be affected by surrounding vehicles. Prediction and decision should not be considered as two separate stages because all vehicles make decisions interactively. This paper constructs the multi-vehicle driving scenario as a non-zero-sum game and proposes a novel game control framework, which consider prediction, decision and control as a whole. The mutual influence of interactions between vehicles is considered in this framework because decisions are made by Nash equilibrium strategy. To efficiently obtain the strategy, ADP, a model-based reinforcement learning method, is used to solve coupled Hamilton-Jacobi-Bellman equations. Driving performance is evaluated by tracking, efficiency, safety and comfort indices. Experiments show that our algorithm could drive perfectly by directly controlling acceleration and steering angle. Vehicles could learn interactive behaviors such as overtaking and pass. In summary, we propose a non-zero-sum game framework for modeling multi-vehicle driving, provide an effective way to solve the Nash equilibrium driving strategy, and validate at non-signalized intersections.
翻译:当车辆在路上驾驶时,其行为将受到周围车辆的影响。预测和决定不应被视为两个不同的阶段,因为所有车辆都以交互方式作出决定。本文将多车辆驾驶方案构建为非零和游戏,并提出一个新的游戏控制框架,将预测、决定和控制作为一个整体加以考虑。在本框架中,车辆相互作用的相互影响会得到考虑,因为决定是由纳什平衡战略作出的。为了高效获得战略,ADP(基于模型的强化学习方法)被用来解决汉密尔顿-Jacobi-Bellman等式的结合问题。驾驶性能通过跟踪、效率、安全和舒适指数进行评估。实验表明,我们的算法可以完美地通过直接控制加速和方向来驱动。车辆可以学习诸如超载和通过等互动行为。总而言之,我们提议了一个模拟多车辆驾驶的非零和游戏框架,提供解决纳什平衡驾驶战略的有效途径,并在非信号交界处验证。