Training on class-imbalanced data usually results in biased models that tend to predict samples into the majority classes, which is a common and notorious problem. From the perspective of energy-based model, we demonstrate that the free energies of categories are aligned with the label distribution theoretically, thus the energies of different classes are expected to be close to each other when aiming for ``balanced'' performance. However, we discover a severe energy-bias phenomenon in the models trained on class-imbalanced dataset. To eliminate the bias, we propose a simple and effective method named Energy Aligning by merely adding the calculated shift scalars onto the output logits during inference, which does not require to (i) modify the network architectures, (ii) intervene the standard learning paradigm, (iii) perform two-stage training. The proposed algorithm is evaluated on two class imbalance-related tasks under various settings: class incremental learning and long-tailed recognition. Experimental results show that energy aligning can effectively alleviate class imbalance issue and outperform state-of-the-art methods on several benchmarks.
翻译:课堂平衡数据培训通常产生偏差模式,倾向于将样本预测到多数阶层,这是一个常见和臭名昭著的问题。从基于能源的模式的角度来看,我们证明,从理论上讲,各类自由能量与标签分配一致,因此,在追求平衡性能时,不同阶层的能量预期会彼此接近。然而,我们发现,在接受过课堂平衡数据集培训的模型中,存在着严重的能源偏差现象。为了消除这种偏差,我们建议一种简单而有效的方法,即能源对齐,在推断过程中,仅仅在产出记录上加上计算好的转换标值,这不需要(一) 修改网络结构,(二) 干预标准学习模式,(三) 进行两阶段培训。在不同的环境下,对与班级不平衡有关的两个任务进行评估:班级递增学习和长尾的识别。实验结果表明,能源对齐能有效减轻阶级不平衡问题,并在几个基准上采用超常规的方法。