Emergency vehicles (EMVs) play a crucial role in responding to time-critical calls such as medical emergencies and fire outbreaks in urban areas. Existing methods for EMV dispatch typically optimize routes based on historical traffic-flow data and design traffic signal pre-emption accordingly; however, we still lack a systematic methodology to address the coupling between EMV routing and traffic signal control. In this paper, we propose EMVLight, a decentralized reinforcement learning (RL) framework for joint dynamic EMV routing and traffic signal pre-emption. We adopt the multi-agent advantage actor-critic method with policy sharing and spatial discounted factor. This framework addresses the coupling between EMV navigation and traffic signal control via an innovative design of multi-class RL agents and a novel pressure-based reward function. The proposed methodology enables EMVLight to learn network-level cooperative traffic signal phasing strategies that not only reduce EMV travel time but also shortens the travel time of non-EMVs. Simulation-based experiments indicate that EMVLight enables up to a $42.6\%$ reduction in EMV travel time as well as an $23.5\%$ shorter average travel time compared with existing approaches.
翻译:紧急车辆(EMVs)在应对医疗紧急情况和城市地区火灾爆发等时间紧迫的呼叫方面发挥着关键作用。现有快速机动车发送方法通常根据历史交通流量数据优化路线,并据此设计交通信号先发制人;然而,我们仍缺乏系统的方法来解决快速机动车路线和交通信号控制之间的交错问题。在本文件中,我们提议为联合动态机动车路线和交通信号先发制人采用分散式强化学习框架(RL),用于联合动态机动车路线和交通信号先发制人。我们采用多剂优势行为者-丙型方法,共享政策和空间折扣系数。这一框架通过多级RL代理的创新设计和新的压力奖励功能,解决了机动车导航和交通信号控制之间的交错。拟议方法使快速机动车能够学习网络级合作交通信号分阶段战略,不仅缩短了快速机动车旅行时间,而且缩短了非机动车的差旅时间。模拟实验表明,EMVLight公司可以将流动旅行时间减少42.6美元,还相当于23.5美元的平均旅行时间。