Ensembles are widely used in machine learning and, usually, provide state-of-the-art performance in many prediction tasks. From the very beginning, the diversity of an ensemble has been identified as a key factor for the superior performance of these models. But the exact role that diversity plays in ensemble models is poorly understood, specially in the context of neural networks. In this work, we combine and expand previously published results in a theoretically sound framework that describes the relationship between diversity and ensemble performance for a wide range of ensemble methods. More precisely, we provide sound answers to the following questions: how to measure diversity, how diversity relates to the generalization error of an ensemble, and how diversity is promoted by neural network ensemble algorithms. This analysis covers three widely used loss functions, namely, the squared loss, the cross-entropy loss, and the 0-1 loss; and two widely used model combination strategies, namely, model averaging and weighted majority vote. We empirically validate this theoretical analysis with neural network ensembles.
翻译:集成在机器学习中广泛使用,通常在很多预测任务中提供最先进的性能。从一开始,组合的多样性就被确定为这些模型优异性能的一个关键因素。但是,多样性在组合模型中的确切作用并没有得到很好理解,特别是在神经网络中。在这项工作中,我们结合并扩展了先前公布的结果,形成了一个理论健全的框架,其中描述了多样性和组合性绩效之间的关系,用于一系列广泛的组合方法。更确切地说,我们为下列问题提供了正确的答案:如何衡量多样性,多样性如何与组合性的一般错误相关,以及神经网络共通算法如何促进多样性。这一分析涵盖了三种广泛使用的损失功能,即平方损失、交叉作物损失和0-1损失;以及两种广泛使用的模型组合战略,即模型平均和加权多数票。我们用神经网络的组合对理论分析进行了经验验证。