The conventional wisdom behind learning deep classification models is to focus on bad-classified examples and ignore well-classified examples that are far from the decision boundary. For instance, when training with cross-entropy loss, examples with higher likelihoods (i.e., well-classified examples) contribute smaller gradients in back-propagation. However, we theoretically show that this common practice hinders representation learning, energy optimization, and margin growth. To counteract this deficiency, we propose to reward well-classified examples with additive bonuses to revive their contribution to the learning process. This counterexample theoretically addresses these three issues. We empirically support this claim by directly verifying the theoretical results or significant performance improvement with our counterexample on diverse tasks, including image classification, graph classification, and machine translation. Furthermore, this paper shows that we can deal with complex scenarios, such as imbalanced classification, OOD detection, and applications under adversarial attacks because our idea can solve these three issues. Code is available at: https://github.com/lancopku/well-classified-examples-are-underestimated.
翻译:学习深层次分类模型的传统智慧是侧重于错误分类实例,忽视远离决定界限的分类实例。例如,在交叉热带损失培训中,可能性较高的实例(例如分类良好的实例)有助于后方宣传中的梯度较小。然而,我们理论上表明,这种常见做法阻碍代表性学习、能源优化和边际增长。为了弥补这一缺陷,我们提议奖励有添加奖金的分类实例,以恢复其对学习进程的贡献。这个反例理论上解决了这三个问题。我们通过直接核查理论结果或显著的绩效改进,与我们关于不同任务的对应实例,包括图像分类、图表分类和机器翻译,从经验上支持这一主张。此外,本文表明,我们可以处理复杂的情景,例如不平衡的分类、OOD探测和在对抗性攻击下的应用,因为我们的想法可以解决这三个问题。代码可在以下网址查阅:https://github.com/lancopku/well-clicle-examples-are-underader。</s>