Dependency networks (Heckerman et al., 2000) are potential probabilistic graphical models for systems comprising a large number of variables. Like Bayesian networks, the structure of a dependency network is represented by a directed graph, and each node has a conditional probability table. Learning and inference are realized locally on individual nodes; therefore, computation remains tractable even with a large number of variables. However, the dependency network's learned distribution is the stationary distribution of a Markov chain called pseudo-Gibbs sampling and has no closed-form expressions. This technical disadvantage has impeded the development of dependency networks. In this paper, we consider a certain manifold for each node. Then, we can interpret pseudo-Gibbs sampling as iterative m-projections onto these manifolds. This interpretation provides a theoretical bound for the location where the stationary distribution of pseudo-Gibbs sampling exists in distribution space. Furthermore, this interpretation involves structure and parameter learning algorithms as optimization problems. In addition, we compare dependency and Bayesian networks experimentally. The results demonstrate that the dependency network and the Bayesian network have roughly the same performance in terms of the accuracy of their learned distributions. The results also show that the dependency network can learn much faster than the Bayesian network.
翻译:依赖性网络(Heckerman等人,2000年)是构成大量变量的系统的潜在概率图形模型(Heckerman等人,2000年)。与巴伊西亚网络一样,依赖性网络的结构以定向图表为代表,每个节点都有一个有条件的概率表。学习和推断在单个节点上就在当地实现;因此,即使有大量变量,计算也仍然是可移植的。但是,依赖性网络所学的分布是一个叫作假吉布斯取样的Markov链的固定分布,没有封闭式表达式。这一技术劣势阻碍了依赖性网络的发展。在本文中,我们认为每个节点都有一定的方块。然后,我们可以将伪吉布斯取样作为这些节点的迭代式模型加以解释。这种解释为在分布空间存在伪吉布斯取样的固定分布位置提供了理论约束。此外,这种解释还涉及结构和参数学习算法作为优化问题。此外,我们比较依赖性和贝亚网络的表达力妨碍了依赖性网络的发展。结果表明,依赖性网络和巴伊西亚网络的分布速度也比学习网络的准确性要快得多。