Learning a Bayesian network (BN) from data can be useful for decision-making or discovering causal relationships. However, traditional methods often fail in modern applications, which exhibit a larger number of observed variables than data points. The resulting uncertainty about the underlying network as well as the desire to incorporate prior information recommend a Bayesian approach to learning the BN, but the highly combinatorial structure of BNs poses a striking challenge for inference. The current state-of-the-art methods such as order MCMC are faster than previous methods but prevent the use of many natural structural priors and still have running time exponential in the maximum indegree of the true directed acyclic graph (DAG) of the BN. We here propose an alternative posterior approximation based on the observation that, if we incorporate empirical conditional independence tests, we can focus on a high-probability DAG associated with each order of the vertices. We show that our method allows the desired flexibility in prior specification, removes timing dependence on the maximum indegree and yields provably good posterior approximations; in addition, we show that it achieves superior accuracy, scalability, and sampler mixing on several datasets.
翻译:从数据中学习巴伊西亚网络(BN)对于决策或发现因果关系可能有用。然而,传统方法往往在现代应用中失败,现代应用中显示的观测变量比数据点多。因此,基础网络的不确定性以及纳入事先信息的愿望建议采用巴伊西亚方法学习巴伊西亚网络,但BN的高度组合结构对推论提出了惊人的挑战。目前最先进的方法,如命令MCMCM比以前的方法要快,但防止了许多自然结构前置物的使用,并且仍然在BN真正定向循环图(DAG)的最大程度上有时间指数化。我们在此提出一个替代的远地点近似值,其依据的观察,即如果我们纳入实验性有条件独立测试,我们就可以把重点放在与每个脊椎顺序相关的高概率DAG上。我们表明,我们的方法允许在事先规格上具有所期望的灵活性,可以消除对最高度和产量良好的近似值的定时依赖;此外,我们还表明,它实现了精确性、可缩度和几度数据的精确性。