Imitation learning is a powerful approach for learning autonomous driving policy by leveraging data from expert driver demonstrations. However, driving policies trained via imitation learning that neglect the causal structure of expert demonstrations yield two undesirable behaviors: inertia and collision. In this paper, we propose Causal Imitative Model (CIM) to address inertia and collision problems. CIM explicitly discovers the causal model and utilizes it to train the policy. Specifically, CIM disentangles the input to a set of latent variables, selects the causal variables, and determines the next position by leveraging the selected variables. Our experiments show that our method outperforms previous work in terms of inertia and collision rates. Moreover, thanks to exploiting the causal structure, CIM shrinks the input dimension to only two, hence, can adapt to new environments in a few-shot setting. Code is available at https://github.com/vita-epfl/CIM.
翻译:模拟学习是利用专家驾驶员演示中的数据学习自主驱动政策的有力方法。然而,通过模拟学习所培训的忽视专家示范的因果结构的驾驶政策产生了两种不可取的行为:惯性与碰撞。在本文中,我们提出“因果模拟模型”以解决惯性与碰撞问题。CIM明确发现因果模型,并利用它来培训政策。具体地说,CIM将输入的内容与一组潜在变量脱钩,选择因果变量,并通过利用选定的变量确定下一个位置。我们的实验表明,我们的方法在惯性与碰撞率方面超过了先前的工作。此外,由于利用因果结构,CIM将输入的层面缩小至仅两个,因此,CIM可以在几眼环境中适应新的环境。代码可在https://github.com/vita-epfl/CIM上查阅。