Learning the structure of a Bayesian Network (BN) with score-based solutions involves exploring the search space of possible graphs and moving towards the graph that maximises a given objective function. Some algorithms offer exact solutions that guarantee to return the graph with the highest objective score, while others offer approximate solutions in exchange for reduced computational complexity. This paper describes an approximate BN structure learning algorithm, which we call Model Averaging Hill-Climbing (MAHC), that combines two novel strategies with hill-climbing search. The algorithm starts by pruning the search space of graphs, where the pruning strategy can be viewed as an aggressive version of the pruning strategies that are typically applied to combinatorial optimisation structure learning problems. It then performs model averaging in the hill-climbing search process and moves to the neighbouring graph that maximises the objective function, on average, for that neighbouring graph and over all its valid neighbouring graphs. Comparisons with other algorithms spanning different classes of learning suggest that the combination of aggressive pruning with model averaging is both effective and efficient, particularly in the presence of data noise.
翻译:学习一个基于分数的巴伊西亚网络(BN) 结构, 包括探索可能的图表的搜索空间, 并移动到能实现给定目标功能最大化的图表。 有些算法提供了精确的解决方案, 保证以最高目标分返回图表, 而另一些算法则提供了近似解决方案, 以降低计算复杂性为交换条件。 本文描述了一种大致的巴伊西亚结构学习算法, 我们称之为模型 Averashing Hill- Climbing( MAHC ), 它将两种新颖策略与山坡搜索相结合。 算法从修剪图的搜索空间开始, 可以将图谱的搜索空间视为典型用于梳理优化结构学习问题的快速化战略版本。 然后, 它将模型平均地用于山坡搜索过程, 并移动到相邻图和所有有效的相邻图形的最大目标功能。 与跨越不同学习类别的其他算法的比较表明, 与模型平均率的侵略性组合是有效和高效的, 特别是在存在数据噪音的情况下。