Deep neural networks generally perform poorly with datasets that suffer from quantity imbalance and classification difficulty imbalance between different classes. In order to alleviate the problem of dataset bias or domain shift in the existing two-stage approaches, a phased progressive learning schedule was proposed for smoothly transferring the training emphasis from representation learning to upper classifier training. This has greater effectivity on datasets that have more severe imbalances or smaller scales. A coupling-regulation-imbalance loss function was designed, coupling a correction term, Focal loss and LDAM loss. Coupling-regulation-imbalance loss can better deal with quantity imbalance and outliers, while regulating focus-of-attention of samples with a variety of classification difficulties. Excellent results were achieved on multiple benchmark datasets using these approaches and they can be easily generalized for other imbalanced classification models. Our code will be open source soon.
翻译:深神经网络一般表现不佳,数据集因数量不平衡和不同类别之间的分类困难而受到影响。为了减轻现有两阶段方法中数据集偏差或域变的问题,建议分阶段逐步学习时间表,将培训重点从代表性学习顺利地从代表制学习转移到高级分类培训,这对具有更严重不平衡或较小比例的数据集具有更大的影响力。设计了一个组合调节-平衡损失功能,将一个修正术语、焦点损失和LDAM损失合并起来。混合调节-平衡损失可以更好地处理数量不平衡和外部值问题,同时以各种分类困难来规范抽样的集中使用。在使用这些方法的多个基准数据集上取得了很好的成果,这些数据集很容易被其他不平衡的分类模型普遍化。我们的代码不久将成为开源。