While neural network binary classifiers are often evaluated on metrics such as Accuracy and $F_1$-Score, they are commonly trained with a cross-entropy objective. How can this training-evaluation gap be addressed? While specific techniques have been adopted to optimize certain confusion matrix based metrics, it is challenging or impossible in some cases to generalize the techniques to other metrics. Adversarial learning approaches have also been proposed to optimize networks via confusion matrix based metrics, but they tend to be much slower than common training methods. In this work, we propose a unifying approach to training neural network binary classifiers that combines a differentiable approximation of the Heaviside function with a probabilistic view of the typical confusion matrix values using soft sets. Our theoretical analysis shows the benefit of using our method to optimize for a given evaluation metric, such as $F_1$-Score, with soft sets, and our extensive experiments show the effectiveness of our approach in several domains.
翻译:虽然神经网络二进制分类器往往根据诸如准确度和1美元-核心等计量标准进行评估,但它们通常都经过跨渗透性目标的培训。 如何解决这种培训-评价差距?虽然已经采用了某些具体技术优化某些基于混杂矩阵的衡量标准,但在某些情况下,将这些技术推广到其他计量标准却具有挑战性或不可能。还提出了反向学习方法,以便通过基于混杂矩阵的计量标准优化网络,但往往比共同的培训方法要慢得多。在这项工作中,我们建议采用一种统一的方法,培训神经网络二进制分类器,将海维赛边功能的不同近似值结合起来,同时对使用软体的典型的混杂矩阵值进行概率分析。我们的理论分析表明,利用我们的方法优化特定评价指标,例如用软体的1美元-核心值,以及我们的广泛实验表明我们在若干领域的方法的有效性。