We present consistent algorithms for multiclass learning with complex performance metrics and constraints, where the objective and constraints are defined by arbitrary functions of the confusion matrix. This setting includes many common performance metrics such as the multiclass G-mean and micro F1-measure, and constraints such as those on the classifier's precision and recall and more recent measures of fairness discrepancy. We give a general framework for designing consistent algorithms for such complex design goals by viewing the learning problem as an optimization problem over the set of feasible confusion matrices. We provide multiple instantiations of our framework under different assumptions on the performance metrics and constraints, and in each case show rates of convergence to the optimal (feasible) classifier (and thus asymptotic consistency). Experiments on a variety of multiclass classification tasks and fairness-constrained problems show that our algorithms compare favorably to the state-of-the-art baselines.
翻译:我们为多级学习提供了一致的算法,具有复杂的业绩计量和制约因素,其中目标和制约因素由混乱矩阵的任意功能来界定,包括许多共同的业绩计量,如多级G平均值和微观F1计量,以及诸如分类员精确度和召回率等制约因素和最近关于公平差异的计量;我们为设计这些复杂设计目标的一致算法提供了一个总体框架,将学习问题视为与一套可行的混乱矩阵相比最优化的问题;我们根据对业绩计量和制约因素的不同假设,提供了我们框架的多重即时反应,在每种情况下都显示与最佳(可行)分类员的趋同率(从而也表明一致性 ) 。 对多种多级分类任务和受到公平制约的问题进行的实验表明,我们的算法与最先进的基线相比是有利的。