We present a sound and complete algorithm, called iterative causal discovery (ICD), for recovering causal graphs in the presence of latent confounders and selection bias. ICD relies on the causal Markov and faithfulness assumptions and recovers the equivalence class of the underlying causal graph. It starts with a complete graph, and consists of a single iterative stage that gradually refines this graph by identifying conditional independence (CI) between connected nodes. Independence and causal relations entailed after any iteration are correct, rendering ICD anytime. Essentially, we tie the size of the CI conditioning set to its distance on the graph from the tested nodes, and increase this value in the successive iteration. Thus, each iteration refines a graph that was recovered by previous iterations having smaller conditioning sets -- a higher statistical power -- which contributes to stability. We demonstrate empirically that ICD requires significantly fewer CI tests and learns more accurate causal graphs compared to FCI, FCI+, and RFCI algorithms.
翻译:我们提出了一个健全和完整的算法,称为迭代因果发现(ICD),用于在潜在混淆者和选择偏差的情况下恢复因果图。 ICD依靠因果马尔科夫和忠诚假设,并恢复基本因果图的等值。它从完整的图表开始,由单一的迭代阶段组成,通过确定连接节点之间的有条件独立(CI)来逐步完善该图。在任何迭代后产生的独立和因果关系是正确的,使ICD随时都是正确的。从根本上说,我们把CI调控装置的大小与图表上与测试节点的距离联系起来,并在连续迭代中增加这一值。因此,每个迭代都精细地细化了由具有更小的因果组合(一种更高的统计能力)的先前迭代数所恢复的图表,从而有助于稳定性。我们从经验上证明, ICD要求远比FCI、FCI+和RCI算法要少得多的因果图。