Causal discovery and causal reasoning are classically treated as separate and consecutive tasks: one first infers the causal graph, and then uses it to estimate causal effects of interventions. However, such a two-stage approach is uneconomical, especially in terms of actively collected interventional data, since the causal query of interest may not require a fully-specified causal model. From a Bayesian perspective, it is also unnatural, since a causal query (e.g., the causal graph or some causal effect) can be viewed as a latent quantity subject to posterior inference -- other unobserved quantities that are not of direct interest (e.g., the full causal model) ought to be marginalized out in this process and contribute to our epistemic uncertainty. In this work, we propose Active Bayesian Causal Inference (ABCI), a fully-Bayesian active learning framework for integrated causal discovery and reasoning, which jointly infers a posterior over causal models and queries of interest. In our approach to ABCI, we focus on the class of causally-sufficient, nonlinear additive noise models, which we model using Gaussian processes. We sequentially design experiments that are maximally informative about our target causal query, collect the corresponding interventional data, and update our beliefs to choose the next experiment. Through simulations, we demonstrate that our approach is more data-efficient than several baselines that only focus on learning the full causal graph. This allows us to accurately learn downstream causal queries from fewer samples while providing well-calibrated uncertainty estimates for the quantities of interest.
翻译:因果关系的发现和因果推理通常被视为单独和连续的任务:首先,先推断因果图,然后使用它来估计干预的因果影响。然而,这种两阶段方法是不经济的,特别是积极收集的干预数据,因为因果查询可能不需要完全指定的因果模型。从巴伊西亚人的角度来看,它也是不自然的,因为因果查询(如因果图或某种因果影响)可被视为一种潜在数量,但需事后推断,其他不直接感兴趣的未观察到的数量(例如,完全因果模型)应在此过程中被排挤出去,并助长我们所发现的因果不确定性。在这项工作中,我们提出“因果查询”可能不需要完全特定的因果模型。从全巴伊西亚的因果发现和推理学框架(ABCI)可以共同推导出一种因果模型,而不是因果关系模型和利息查询。在我们对ABCI的处理方法中,我们注重的因果、非因果性、非因果的因果模型应该被排入的量数量,我们选择了一系列的因果实验数据,我们用最低的实验方法来采集数据。