The target of reducing travel time only is insufficient to support the development of future smart transportation systems. To align with the United Nations Sustainable Development Goals (UN-SDG), a further reduction of fuel and emissions, improvements of traffic safety, and the ease of infrastructure deployment and maintenance should also be considered. Different from existing work focusing on the optimization of the control in either traffic light signal (to improve the intersection throughput), or vehicle speed (to stabilize the traffic), this paper presents a multi-agent Deep Reinforcement Learning (DRL) system called CoTV, which Cooperatively controls both Traffic light signals and Connected Autonomous Vehicles (CAV). Therefore, our CoTV can well balance the achievement of the reduction of travel time, fuel, and emissions. In the meantime, CoTV can also be easy to deploy by cooperating with only one CAV that is the nearest to the traffic light controller on each incoming road. This enables more efficient coordination between traffic light controllers and CAV, thus leading to the convergence of training CoTV under the large-scale multi-agent scenario that is traditionally difficult to converge. We give the detailed system design of CoTV and demonstrate its effectiveness in a simulation study using SUMO under various grid maps and realistic urban scenarios with mixed-autonomy traffic.
翻译:为了与联合国可持续发展目标(UN-SDG)保持一致,还应考虑进一步减少燃料和排放、改善交通安全以及便利基础设施的部署和维护。与目前侧重于优化对交通灯信号(以改进交汇路面)或车辆速度(稳定交通)的控制的工作不同,本文件提出了一个多试管深强化学习系统(DRL),称为“COTV”,该系统共同控制交通灯信号和连接自动车辆(CAV)。因此,我们的COTV可以很好地平衡减少交通时间、燃料和排放的实现。与此同时,CTV也容易部署,只与靠近每条行进公路交通灯控制器的一台CAV合作,从而能够更有效地协调交通灯控制器和CAV,从而使得CTV培训在传统上难以合并的大型多试样情景下趋于一致。我们提供了COTV的详细系统设计,并在使用各种电网图和混合城市地图的模拟情景中展示其模拟交通的有效性。