The robustness of any machine learning solution is fundamentally bound by the data it was trained on. One way to generalize beyond the original training is through human-informed augmentation of the original dataset; however, it is impossible to specify all possible failure cases that can occur during deployment. To address this limitation we combine model-based reinforcement learning and model-interpretability methods to propose a solution that self-generates simulated scenarios constrained by environmental concepts and dynamics learned in an unsupervised manner. In particular, an internal model of the agent's environment is conditioned on low-dimensional concept representations of the input space that are sensitive to the agent's actions. We demonstrate this method within a standard realistic driving simulator in a simple point-to-point navigation task, where we show dramatic improvements in one-shot generalization to different instances of specified failure cases as well as zero-shot generalization to similar variations compared to model-based and model-free approaches.
翻译:任何机器学习解决方案的稳健性从根本上受它所培训的数据的约束。超出原始培训范围的一种普及方式是通过人为知情地增强原始数据集;然而,不可能具体说明部署期间可能出现的所有可能的故障案例。为了应对这一局限性,我们将基于模型的强化学习和模型解释方法结合起来,提出一种解决方案,使受环境概念和动态以不受监督的方式学习而受环境概念和动态制约的模拟情景自我生成。特别是,该剂环境的内部模型取决于对该剂行动敏感的输入空间的低维概念表达。我们用标准的现实驾驶模拟器在简单点对点导航任务中演示了这种方法,我们展示了这一方法,在对特定故障案例的不同实例的一枪式概括化方面,以及相对于基于模型和无模型方法的类似变化的零弹式概括化。