Classifiers trained with class-imbalanced data are known to perform poorly on test data of the "minor" classes, of which we have insufficient training data. In this paper, we investigate learning a ConvNet classifier under such a scenario. We found that a ConvNet significantly over-fits the minor classes, which is quite opposite to traditional machine learning algorithms that often under-fit minor classes. We conducted a series of analysis and discovered the feature deviation phenomenon -- the learned ConvNet generates deviated features between the training and test data of minor classes -- which explains how over-fitting happens. To compensate for the effect of feature deviation which pushes test data toward low decision value regions, we propose to incorporate class-dependent temperatures (CDT) in training a ConvNet. CDT simulates feature deviation in the training phase, forcing the ConvNet to enlarge the decision values for minor-class data so that it can overcome real feature deviation in the test phase. We validate our approach on benchmark datasets and achieve promising performance. We hope that our insights can inspire new ways of thinking in resolving class-imbalanced deep learning.
翻译:受过班级平衡数据培训的分类者在“初级”班的测试数据上表现不佳,我们没有足够的培训数据。在本文中,我们调查在这种情景下学习ConvNet分类器的情况。我们发现ConvNet大大超出了次要班级,这与传统的机器学习算法大相径庭,而传统的机器学习算法往往不适应次要班级。我们进行了一系列分析并发现了特征偏差现象 -- -- 所学的ConvNet在培训和测试小班级数据之间产生偏差特征 -- -- 这解释了如何出现过大的情况。为了弥补将测试数据推向低决策值区域的特征偏差的影响,我们提议在培训ConvNet时纳入依赖班级的温度(CDT)。CDT模拟了培训阶段的偏差,迫使ConvNet扩大次要班级数据的决策值,以便克服测试阶段的真正特征偏差。我们验证了我们关于基准数据集的方法,并取得了有希望的业绩。我们希望我们的洞察能够激发新的思维方法,解决班级平衡的深层次学习。