In this paper, we propose a balancing training method to address problems in imbalanced data learning. To this end, we derive a new loss used in the balancing training phase that alleviates the influence of samples that cause an overfitted decision boundary. The proposed loss efficiently improves the performance of any type of imbalance learning methods. In experiments on multiple benchmark data sets, we demonstrate the validity of our method and reveal that the proposed loss outperforms the state-of-the-art cost-sensitive loss methods. Furthermore, since our loss is not restricted to a specific task, model, or training method, it can be easily used in combination with other recent re-sampling, meta-learning, and cost-sensitive learning methods for class-imbalance problems.
翻译:在本文中,我们提出了一种平衡培训方法,以解决数据学习不平衡的问题。为此目的,我们在平衡培训阶段中发现了一种新的损失,减轻了造成决定界限过宽的样本的影响。拟议的损失有效地改善了任何类型不平衡学习方法的绩效。在多个基准数据集的实验中,我们展示了我们的方法的有效性,并揭示了拟议的损失超过了最先进的成本敏感损失方法。此外,由于我们的损失并不局限于特定的任务、模式或培训方法,因此很容易与其他近期的重新采样、元学习和成本敏感的学习方法结合使用,以解决课堂平衡问题。