This paper proposes an eco-driving framework for electric connected vehicles (CVs) based on reinforcement learning (RL) to improve vehicle energy efficiency at signalized intersections. The vehicle agent is specified by integrating the model-based car-following policy, lane-changing policy, and the RL policy, to ensure safe operation of a CV. Subsequently, a Markov Decision Process (MDP) is formulated, which enables the vehicle to perform longitudinal control and lateral decisions, jointly optimizing the car-following and lane-changing behaviors of the CVs in the vicinity of intersections. Then, the hybrid action space is parameterized as a hierarchical structure and thereby trains the agents with two-dimensional motion patterns in a dynamic traffic environment. Finally, our proposed methods are evaluated in SUMO software from both a single-vehicle-based perspective and a flow-based perspective. The results show that our strategy can significantly reduce energy consumption by learning proper action schemes without any interruption of other human-driven vehicles (HDVs).
翻译:本文根据强化学习(RL),提出了电动联动车辆生态驱动框架,以提高信号十字路口的车辆能效。车辆代理器通过整合基于模式的汽车跟踪政策、车道改变政策和车辆保动政策,确保车辆安全运行。随后,制定了Markov决策程序(MDP),使车辆能够进行纵向控制和横向决定,共同优化十字路口附近的汽车跟踪和车道改变行为。随后,混合行动空间被标为等级结构,从而在动态交通环境中用二维运动模式对车辆进行训练。最后,我们提出的方法从单一车辆和流动角度在苏莫软件中进行评估。结果显示,我们的战略可以大大减少能源消耗,方法是学习适当的行动计划,而不会干扰其他人类驱动的车辆(HDVs)。