Ensemble Learning methods combine multiple algorithms performing the same task to build a group with superior quality. These systems are well adapted to the distributed setup, where each peer or machine of the network hosts one algorithm and communicate its results to its peers. Ensemble learning methods are naturally resilient to the absence of several peers thanks to the ensemble redundancy. However, the network can be corrupted, altering the prediction accuracy of a peer, which has a deleterious effect on the ensemble quality. In this paper, we propose a noise-resilient ensemble classification method, which helps to improve accuracy and correct random errors. The approach is inspired by Evidence Accumulation Clustering , adapted to classification ensembles. We compared it to the naive voter model over four multi-class datasets. Our model showed a greater resilience, allowing us to recover prediction under a very high noise level. In addition as the method is based on the evidence accumulation clustering, our method is highly flexible as it can combines classifiers with different label definitions.
翻译:组合式学习方法结合了多种算法,这些算法执行同样的任务,以构建质量更高的群体。 这些系统非常适合分布式设置, 网络的每个同侪或机器都使用一个算法, 并将结果传达给同侪。 组合式学习方法自然具有适应性, 因为由于组合式冗余, 多个同侪没有。 然而, 网络可能会被损坏, 改变同侪的预测准确性, 从而改变同侪的预测准确性, 从而对同龄人的质量产生有害影响 。 在本文中, 我们提出一种静默的混合分类方法, 有助于改进准确性和纠正随机错误 。 这种方法受证据累积组合组合的启发, 并适应分类组合式 。 我们将其与四个多级数据集的天真选民模型进行比较 。 我们的模型显示更大的弹性, 使我们能够在非常高的噪音水平下恢复预测 。 此外, 由于方法基于证据累积组合, 我们的方法非常灵活, 因为它可以将分类和不同标签定义结合起来 。