When minimizing the empirical risk in binary classification, it is a common practice to replace the zero-one loss with a surrogate loss to make the learning objective feasible to optimize. Examples of well-known surrogate losses for binary classification include the logistic loss, hinge loss, and sigmoid loss. It is known that the choice of a surrogate loss can highly influence the performance of the trained classifier and therefore it should be carefully chosen. Recently, surrogate losses that satisfy a certain symmetric condition (aka., symmetric losses) have demonstrated their usefulness in learning from corrupted labels. In this article, we provide an overview of symmetric losses and their applications. First, we review how a symmetric loss can yield robust classification from corrupted labels in balanced error rate (BER) minimization and area under the receiver operating characteristic curve (AUC) maximization. Then, we demonstrate how the robust AUC maximization method can benefit natural language processing in the problem where we want to learn only from relevant keywords and unlabeled documents. Finally, we conclude this article by discussing future directions, including potential applications of symmetric losses for reliable machine learning and the design of non-symmetric losses that can benefit from the symmetric condition.
翻译:在二进制分类中,将经验风险降到最低时,通常的做法是用代位损失取代零一损失,使学习目标达到最佳优化。二进制分类中众所周知的代位损失的例子包括后勤损失、断链损失和类固醇损失。已知代位损失的选择会大大影响经过训练的分类员的性能,因此应仔细选择。最近,符合某种对称条件(aka.,对称损失)的代位损失在学习腐败标签方面显示了其有用性。在本篇文章中,我们概述了对称损失及其应用。首先,我们审查了对称损失如何从平衡的误差率(BER)最小化和接收器特性曲线(AUSC)下失密地区的失差率中产生稳健的分类。然后,我们展示了强健的ACUC最大化方法如何在问题中有益于自然语言处理,因为我们希望从相关关键词和未加标签的文件中学习。最后,我们通过讨论对准损失的今后方向,包括从可靠的设计损失中学习可靠的设计结果,从而获得可靠的设计损失的潜在收益。