Autonomous mobility-on-demand (AMoD) systems represent a rapidly developing mode of transportation wherein travel requests are dynamically handled by a coordinated fleet of robotic, self-driving vehicles. Given a graph representation of the transportation network - one where, for example, nodes represent areas of the city, and edges the connectivity between them - we argue that the AMoD control problem is naturally cast as a node-wise decision-making problem. In this paper, we propose a deep reinforcement learning framework to control the rebalancing of AMoD systems through graph neural networks. Crucially, we demonstrate that graph neural networks enable reinforcement learning agents to recover behavior policies that are significantly more transferable, generalizable, and scalable than policies learned through other approaches. Empirically, we show how the learned policies exhibit promising zero-shot transfer capabilities when faced with critical portability tasks such as inter-city generalization, service area expansion, and adaptation to potentially complex urban topologies.
翻译:自动按需流动(AMOD)系统代表着一种迅速发展的运输方式,其中旅行申请由一组协调的机器人驾驶车辆来动态处理。鉴于运输网络的图表显示,例如节点代表城市地区,并超越了两者之间的连接,我们认为,自动按需流动(AMOD)系统自然会成为一个节点决策问题。在本文件中,我们提议了一个深度强化学习框架,以控制通过图形神经网络对AMOD系统进行再平衡。关键是,我们通过图形神经网络显示,图形神经网络使强化学习剂能够恢复比其他方法所学到的政策更可转让、可普及和可扩展的行为政策。我们经常地展示,在面对关键可转移任务时,如城市间一般化、服务领域扩展和适应潜在复杂的城市地形时,我们所学的政策展示了有希望的零速转移能力。