Classification is one of the most studied tasks in data mining and machine learning areas and many works in the literature have been presented to solve classification problems for multiple fields of knowledge such as medicine, biology, security, and remote sensing. Since there is no single classifier that achieves the best results for all kinds of applications, a good alternative is to adopt classifier fusion strategies. A key point in the success of classifier fusion approaches is the combination of diversity and accuracy among classifiers belonging to an ensemble. With a large amount of classification models available in the literature, one challenge is the choice of the most suitable classifiers to compose the final classification system, which generates the need of classifier selection strategies. We address this point by proposing a framework for classifier selection and fusion based on a four-step protocol called CIF-E (Classifiers, Initialization, Fitness function, and Evolutionary algorithm). We implement and evaluate 24 varied ensemble approaches following the proposed CIF-E protocol and we are able to find the most accurate approach. A comparative analysis has also been performed among the best approaches and many other baselines from the literature. The experiments show that the proposed evolutionary approach based on Univariate Marginal Distribution Algorithm (UMDA) can outperform the state-of-the-art literature approaches in many well-known UCI datasets.
翻译:分类是数据挖掘和机器学习领域研究最多的任务之一,文献中的许多著作都提出了解决医学、生物学、安全和遥感等多种知识领域分类问题的最适当分类方法,因为没有一个单一的分类师能够取得所有各种应用的最佳结果,因此,一个好的替代办法是采用分类融合战略。分类融合方法取得成功的一个关键点是属于一个共同体的分类师之间多样性和准确性相结合。由于文献中有大量分类模型,一个挑战是选择最合适的分类师来组成最终分类系统,从而产生分类员选择战略的需要。我们通过提出一个分类员选择和融合框架来解决这个问题,该框架的基础是四步协议,即CIF-E(分类员、初始化、完善功能和进化算法)。我们实施和评价了拟议的CIF-E协议的24种混合方法,我们能够找到最准确的方法。在文献中的最佳方法和许多其他基线中也进行了比较分析。Universal-IFA-S-MM-S-Silvarial-Silvarial-Silvarial-Anial-Alivorial-Alistria-ILAlistria-Alistal-IFAlistal-IFAdal-IG-IG-IFF MI-IMA-S-ILAdal-ILAdal-S-IFF-S-IMA-IMA-IMA-IMA-S-S-IMA-S-S-IFF-S-S-S-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-IMA-MA-MA-MA-MA-MA-MA-MA-MA-MA-