In the autonomous driving field, fusion of human knowledge into Deep Reinforcement Learning (DRL) is often based on the human demonstration recorded in a simulated environment. This limits the generalization and the feasibility of application in real-world traffic. We propose a two-stage DRL method to train a car-following agent, that modifies the policy by leveraging the real-world human driving experience and achieves performance superior to the pure DRL agent. Training a DRL agent is done within CARLA framework with Robot Operating System (ROS). For evaluation, we designed different driving scenarios to compare the proposed two-stage DRL car-following agent with other agents. After extracting the "good" behavior from the human driver, the agent becomes more efficient and reasonable, which makes this autonomous agent more suitable for Human-Robot Interaction (HRI) traffic.
翻译:在自主驾驶领域,将人类知识融入深强化学习(DRL)往往基于模拟环境中记录的人类演示。这限制了在现实世界交通中应用人类知识的普及性和可行性。我们建议采用两阶段DRL方法培训汽车跟踪剂,通过利用现实世界人类驾驶经验来改变政策,并取得优于纯DRL剂的性能。在CARLA框架内,用机器人操作系统(ROS)培训DRL代理剂。为了评估,我们设计了不同的驾驶方案,将拟议的两阶段DRL汽车跟踪剂与其他代理器进行比较。在从人类司机身上提取“良好”行为之后,该代理器变得更加高效和合理,从而使这一自主代理器更适合人类机器人的互动。