Recently, several studies have claimed that using class-specific feature subsets provides certain advantages over using a single feature subset for representing the data for a classification problem. Unlike traditional feature selection methods, the class-specific feature selection methods select an optimal feature subset for each class. Typically class-specific feature selection (CSFS) methods use one-versus-all split of the data set that leads to issues such as class imbalance, decision aggregation, and high computational overhead. We propose a class-specific feature selection method embedded in a fuzzy rule-based classifier, which is free from the drawbacks associated with most existing class-specific methods. Additionally, our method can be adapted to control the level of redundancy in the class-specific feature subsets by adding a suitable regularizer to the learning objective. Our method results in class-specific rules involving class-specific subsets. We also propose an extension where different rules of a particular class are defined by different feature subsets to model different substructures within the class. The effectiveness of the proposed method has been validated through experiments on three synthetic data sets.
翻译:最近,有几项研究声称,使用特定类别特性子集比使用单一特性子集来代表分类问题的数据有一定的优势。与传统的特征选择方法不同,特定类别特征选择方法为每个类别选择一个最佳特性子集。典型的,特定类别特征选择方法使用一对一的数据集分割,导致诸如类别不平衡、决定汇总和高计算间接费用等问题。我们提议了一种特定类别特征选择方法,该方法嵌入一个模糊的基于规则的分类器中,该分类器不附带与大多数现有特定类别方法相关的缺陷。此外,我们的方法可以通过在学习目标中增加一个合适的常规化器来控制特定类别特征子集的冗余程度。我们的方法在涉及特定类别子集的类别特定规则中产生结果。我们还提议扩大一个范围,即特定类别的不同规则由不同的特性子集来界定,以模拟该类的不同子集,以模拟不同结构。通过对三个合成数据集的实验,验证了拟议方法的有效性。