Classifiers have been widely implemented in practice, while how to evaluate them properly remains a problem. Commonly used two types of metrics respectively based on confusion matrix and loss function have different advantages in flexibility and mathematical completeness, while they struggle in different dilemmas like the insensitivity to slight improvements or the lack of customizability in different tasks. In this paper, we propose a novel metric named Meta Pattern Concern Score based on the abstract representation of the probabilistic prediction, as well as the targeted design for processing negative classes in multi-classification and reducing the discreteness of metric value, to achieve advantages of both the two kinds of metrics and avoid their weaknesses. Our metric provides customizability to pick out the model for specific requirements in different practices, and make sure it is also fine under traditional metrics at the same time. Evaluation in four kinds of models and six datasets demonstrates the effectiveness and efficiency of our metric, and a case study shows it can select a model to reduce 0.53% of dangerous misclassifications by sacrificing only 0.04% of training accuracy.
翻译:分类方法在实践中得到了广泛实施,而如何正确评估它们仍然是一个问题。通常使用基于混乱矩阵和损失函数的两种标准在灵活性和数学完整性方面都具有不同优势,而它们在不同的困境中挣扎,如对细微改进不敏感或不同任务缺乏定制性等。在本文中,我们根据概率预测的抽象表述,提出了名为“元模式关注分数”的新指标,以及处理多分类中的负面分类和减少指标值离散性的定向设计,以实现两种指标的优势并避免其缺陷。我们的衡量标准提供了为不同做法的具体要求选择模式的自定义性,并确保在传统指标下也在同一时间很好。四种模型和六个数据集的评估显示了我们指标的有效性和效率,一项案例研究表明,它可以选择一种模式,通过仅牺牲0.04%的培训准确性来减少0.53%的危险分类错误。