Recent work has shown that the introduction of autonomous vehicles (AVs) in traffic could help reduce traffic jams. Deep reinforcement learning methods demonstrate good performance in complex control problems, including autonomous vehicle control, and have been used in state-of-the-art AV controllers. However, deep neural networks (DNNs) render automated driving vulnerable to machine learning-based attacks. In this work, we explore the backdooring/trojanning of DRL-based AV controllers. We develop a trigger design methodology that is based on well-established principles of traffic physics. The malicious actions include vehicle deceleration and acceleration to cause stop-and-go traffic waves to emerge (congestion attacks) or AV acceleration resulting in the AV crashing into the vehicle in front (insurance attack). We test our attack on single-lane and two-lane circuits. Our experimental results show that the backdoored model does not compromise normal operation performance, with the maximum decrease in cumulative rewards being 1%. Still, it can be maliciously activated to cause a crash or congestion when the corresponding triggers appear.
翻译:最近的工作表明,在交通中引入自动车辆有助于减少交通堵塞。深度强化学习方法显示在复杂的控制问题中表现良好,包括自主车辆控制,并被用于最先进的AV控制器。然而,深神经网络使得自动驾驶容易受到机器学习式袭击。在这项工作中,我们探索基于DRL的AV控制器的后门/车道操作。我们开发了一种基于交通物理学既定原则的触发设计方法。恶意行动包括车辆减速和加速导致中途交通波的出现(摄取攻击)或AV加速导致AV在前面撞车(保险攻击)的AV加速。我们测试了单行和双行线路的电路。我们的实验结果表明,后门模式不会损害正常操作性能,而累积收益的最大减幅为1%。但是,当相应的触发器出现时,它可能恶意地引发碰撞或拥堵。