Training self-driving cars is often challenging since they require a vast amount of labeled data in multiple real-world contexts, which is computationally and memory intensive. Researchers often resort to driving simulators to train the agent and transfer the knowledge to a real-world setting. Since simulators lack realistic behavior, these methods are quite inefficient. To address this issue, we introduce a framework (perception, planning, and control) in a real-world driving environment that transfers the real-world environments into gaming environments by setting up a reliable Markov Decision Process (MDP). We propose variations of existing Reinforcement Learning (RL) algorithms in a multi-agent setting to learn and execute the discrete control in real-world environments. Experiments show that the multi-agent setting outperforms the single-agent setting in all the scenarios. We also propose reliable initialization, data augmentation, and training techniques that enable the agents to learn and generalize to navigate in a real-world environment with minimal input video data, and with minimal training. Additionally, to show the efficacy of our proposed algorithm, we deploy our method in the virtual driving environment TORCS.
翻译:培训自我驾驶的汽车往往具有挑战性,因为在多种现实环境中需要大量的标签数据,这是计算和记忆密集的。研究人员常常利用驾驶模拟器来训练代理人并将知识转移到现实环境中。由于模拟器缺乏现实行为,这些方法相当低效。为了解决这个问题,我们在一个真实的驱动环境中引入一个框架(感知、规划和控制),通过建立可靠的马尔科夫决策程序,将真实世界环境转移到游戏环境中。我们提议在多试器环境中修改现有的强化学习算法(RL),以学习和执行现实世界环境中的离散控制。实验显示,在各种情景中,多试剂的设置优于单一代理人设置。我们还提出可靠的初始化、数据增强和培训技术,使代理人能够学习和普及在真实世界环境中的导航,同时提供最低限度的输入视频数据,并进行最低限度的培训。此外,为了显示我们提议的算法的效率,我们把方法运用在虚拟驱动环境中。