While neural network binary classifiers are often evaluated on metrics such as Accuracy and $F_1$-Score, they are commonly trained with a cross-entropy objective. How can this training-testing gap be addressed? While specific techniques have been adopted to optimize certain confusion matrix based metrics, it is challenging or impossible in some cases to generalize the techniques to other metrics. Adversarial learning approaches have also been proposed to optimize networks via confusion matrix based metrics, but they tend to be much slower than common training methods. In this work, we propose to approximate the Heaviside step function, typically used to compute confusion matrix based metrics, to render these metrics amenable to gradient descent. Our extensive experiments show the effectiveness of our end-to-end approach for binary classification in several domains.
翻译:虽然神经网络二元分类器往往根据精确度和1美元-核心等计量标准进行评估,但它们通常受到跨热带目标的培训。如何解决这种培训测试差距?虽然采用了某些具体技术优化某些基于混乱矩阵的计量,但在某些情况下将这些技术推广到其他计量标准是具有挑战性或不可能的。还提出了反向学习方法,通过基于混乱矩阵的计量标准优化网络,但往往比共同的培训方法要慢得多。在这项工作中,我们建议接近 Heaviside 阶梯函数,通常用来计算基于混杂矩阵的计量标准,使这些计量标准易于梯度下降。我们的广泛实验表明,我们在若干领域采用端对端方法进行二元分类是有效的。