Most causal discovery algorithms find causal structure among a set of observed variables. Learning the causal structure among latent variables remains an important open problem, particularly when using high-dimensional data. In this paper, we address a problem for which it is known that inputs cause outputs, and these causal relationships are encoded by a causal network among a set of an unknown number of latent variables. We developed a deep learning model, which we call a redundant input neural network (RINN), with a modified architecture and a regularized objective function to find causal relationships between input, hidden, and output variables. More specifically, our model allows input variables to directly interact with all latent variables in a neural network to influence what information the latent variables should encode in order to generate the output variables accurately. In this setting, the direct connections between input and latent variables makes the latent variables partially interpretable; furthermore, the connectivity among the latent variables in the neural network serves to model their potential causal relationships to each other and to the output variables. A series of simulation experiments provide support that the RINN method can successfully recover latent causal structure between input and output variables.
翻译:多数因果发现算法在一组观察到的变量中找到因果结构。 学习潜伏变量之间的因果结构仍然是一个重要的开放问题, 特别是在使用高维数据时。 在本文中, 我们处理一个已知投入导致产出的问题, 这些因果关系是由一组未知的潜在变量组成的因果网络编码的。 我们开发了一个深层次学习模型, 我们称之为冗余输入神经网络( IRIN), 其结构经过修改, 以及一个常规目标功能, 以寻找输入、 隐藏和输出变量之间的因果关系。 更具体地说, 我们的模式允许输入变量与神经网络中所有潜在变量直接互动, 以影响潜在变量应编码的信息, 以便准确生成输出变量。 在此设置中, 输入和潜在变量之间的直接关联使得潜在变量可以部分解释; 此外, 神经网络中的潜在变量之间的连接有助于模拟它们与输入变量和输出变量之间的潜在因果关系。 一系列模拟实验支持 RINN 方法能够成功恢复输入和输出变量之间的潜在因果结构。