Inducing causal relationships from observations is a classic problem in machine learning. Most work in causality starts from the premise that the causal variables themselves are observed. However, for AI agents such as robots trying to make sense of their environment, the only observables are low-level variables like pixels in images. To generalize well, an agent must induce high-level variables, particularly those which are causal or are affected by causal variables. A central goal for AI and causality is thus the joint discovery of abstract representations and causal structure. However, we note that existing environments for studying causal induction are poorly suited for this objective because they have complicated task-specific causal graphs which are impossible to manipulate parametrically (e.g., number of nodes, sparsity, causal chain length, etc.). In this work, our goal is to facilitate research in learning representations of high-level variables as well as causal structures among them. In order to systematically probe the ability of methods to identify these variables and structures, we design a suite of benchmarking RL environments. We evaluate various representation learning algorithms from the literature and find that explicitly incorporating structure and modularity in models can help causal induction in model-based reinforcement learning.
翻译:从观察中推导因果关系是一个典型的机器学习问题。 大部分因果关系工作都源于因果变数本身被观察到的前提。 但是,对于像机器人这样的AI代理商来说,唯一可见的是像像图像像素这样的低层次变数。 要概括一下,一个代理商必须诱导高层次变数,特别是因果变数或受因果变数影响的变数。因此,AI和因果变数的中心目标是共同发现抽象的表示和因果结构。然而,我们注意到,现有的因果诱导研究环境不适合这一目标,因为它们有复杂的任务特定因果图,无法进行分辨性操作(例如节点的数目、松散、因果链长度等)。 在这项工作中,我们的目标是促进研究高层次变数的表述以及因果结构。为了系统地探索确定这些变数和结构的方法的能力,我们设计一套衡量RL环境的基准环境。我们从文献中评估了各种代表制的算法,发现在模型中明确纳入结构和模块可以帮助加强因果性诱导。