In many supervised learning tasks, the entities to be labeled are related to each other in complex ways and their labels are not independent. For example, in hypertext classification, the labels of linked pages are highly correlated. A standard approach is to classify each entity independently, ignoring the correlations between them. Recently, Probabilistic Relational Models, a relational version of Bayesian networks, were used to define a joint probabilistic model for a collection of related entities. In this paper, we present an alternative framework that builds on (conditional) Markov networks and addresses two limitations of the previous approach. First, undirected models do not impose the acyclicity constraint that hinders representation of many important relational dependencies in directed models. Second, undirected models are well suited for discriminative training, where we optimize the conditional likelihood of the labels given the features, which generally improves classification accuracy. We show how to train these models effectively, and how to use approximate probabilistic inference over the learned model for collective classification of multiple related entities. We provide experimental results on a webpage classification task, showing that accuracy can be significantly improved by modeling relational dependencies.
翻译:在许多受监督的学习任务中,要贴标签的实体以复杂的方式彼此关联,其标签不独立。例如,在超文本分类中,链接页面的标签高度相关。一种标准的做法是将每个实体独立分类,忽视它们之间的相互关系。最近,利用Bayesian网络的关联版概率性关系模型(Bayesian网络的一个关联版本)来定义一个相关实体集合的共同概率模型。在本文件中,我们提出了一个替代框架,以(有条件的)Markov网络为基础,并解决前一种方法的两个局限性。首先,非指令性模型并不强制实行周期性限制,从而阻碍定向模型中许多重要关系依赖性的代表性。第二,非指令性模型非常适合进行歧视培训,我们优化标签的有条件可能性,从而普遍提高分类的准确性。我们展示了如何有效培训这些模型,以及如何在多相关实体的集体分类中使用所学模型的近似概率性推论。我们在网页分类任务中提供实验性结果,显示精确性可以通过大幅改进可靠性。