Bayesian Networks (BNs) represent conditional probability relations among a set of random variables (nodes) in the form of a directed acyclic graph (DAG), and have found diverse applications in knowledge discovery. We study the problem of learning the sparse DAG structure of a BN from continuous observational data. The central problem can be modeled as a mixed-integer program with an objective function composed of a convex quadratic loss function and a regularization penalty subject to linear constraints. The optimal solution to this mathematical program is known to have desirable statistical properties under certain conditions. However, the state-of-the-art optimization solvers are not able to obtain provably optimal solutions to the existing mathematical formulations for medium-size problems within reasonable computational times. To address this difficulty, we tackle the problem from both computational and statistical perspectives. On the one hand, we propose a concrete early stopping criterion to terminate the branch-and-bound process in order to obtain a near-optimal solution to the mixed-integer program, and establish the consistency of this approximate solution. On the other hand, we improve the existing formulations by replacing the linear "big-$M$" constraints that represent the relationship between the continuous and binary indicator variables with second-order conic constraints. Our numerical results demonstrate the effectiveness of the proposed approaches.
翻译:Bayesian Networks(BN) 代表一组随机变量(节点)之间的有条件概率关系,以定向循环图(DAG)为形式,以定向循环图(DAG)为形式的随机变量(节点)为有条件的概率关系,并在知识发现过程中发现多种应用。我们研究从连续观测数据中学习BN的稀疏DAG结构的问题。中心问题可以模拟成混合整数程序,其客观功能由锥形二次损失函数和受线性限制的正规化惩罚组成。已知该数学方案的最佳解决方案在某些条件下具有理想的统计属性。然而,最先进的优化解决方案无法在合理的计算时间内找到现有中等规模问题数学公式的最佳解决方案。为了解决这一困难,我们从计算和统计角度处理问题。一方面,我们提出一个具体的早期停止标准,以终止分支和约束进程,从而获得一种近乎最佳的混合内分数方案解决方案,并确立这一近似的解决办法的一致性。另一方面,我们改进了现有组合解决方案的最佳解决方案,在合理的计算方法之间,我们改进了现有的数学公式,以取代了我们所拟的线性指标上的限制。