This paper addresses the problem of learning a sparse structure Bayesian network from high-dimensional discrete data. Compared to continuous Bayesian networks, learning a discrete Bayesian network is a challenging problem due to the large parameter space. Although many approaches have been developed for learning continuous Bayesian networks, few approaches have been proposed for the discrete ones. In this paper, we address learning Bayesian networks as an optimization problem and propose a score function that satisfies the sparsity and the DAG property simultaneously. Besides, we implement a block-wised stochastic coordinate descent algorithm to optimize the score function. Specifically, we use a variance reducing method in our optimization algorithm to make the algorithm work efficiently in high-dimensional data. The proposed approach is applied to synthetic data from well-known benchmark networks. The quality, scalability, and robustness of the constructed network are measured. Compared to some competitive approaches, the results reveal that our algorithm outperforms the others in evaluation metrics.
翻译:本文探讨从高维离散数据中学习稀疏结构的巴伊西亚网络的问题。 与连续的巴伊西亚网络相比, 学习离散的巴伊西亚网络由于参数空间大而是一个具有挑战性的问题。 虽然已经为学习连续的巴伊西亚网络制定了许多方法, 但很少为离散网络提出办法。 在本文中, 我们把学习巴伊西亚网络作为一个优化问题处理, 并提议一个能同时满足聚度和DAG属性的得分函数。 此外, 我们实施一个有条不紊的随机协调下限算法来优化得分函数。 具体地说, 我们使用一种减少差异的方法来使算法在高维数据中有效发挥作用。 所提议的方法被用于来自众所周知的基准网络的合成数据。 测量了所建网络的质量、 可扩展性和稳健性。 与一些竞争性方法相比, 结果表明, 我们的算法比评价指标中的其他方法要优于其他方法。