Simplicity Bias (SB) is a phenomenon that deep neural networks tend to rely favorably on simpler predictive patterns but ignore some complex features when applied to supervised discriminative tasks. In this work, we investigate SB in long-tailed image recognition and find the tail classes suffer more severely from SB, which harms the generalization performance of such underrepresented classes. We empirically report that self-supervised learning (SSL) can mitigate SB and perform in complementary to the supervised counterpart by enriching the features extracted from tail samples and consequently taking better advantage of such rare samples. However, standard SSL methods are designed without explicitly considering the inherent data distribution in terms of classes and may not be optimal for long-tailed distributed data. To address this limitation, we propose a novel SSL method tailored to imbalanced data. It leverages SSL by triple diverse levels, i.e., holistic-, partial-, and augmented-level, to enhance the learning of predictive complex patterns, which provides the potential to overcome the severe SB on tail data. Both quantitative and qualitative experimental results on five long-tailed benchmark datasets show our method can effectively mitigate SB and significantly outperform the competing state-of-the-arts.
翻译:简易比亚斯(SB)是一种现象,深神经网络往往倾向于依赖更简单的预测模式,但在应用到受监督的歧视性任务时忽视一些复杂的特征。在这项工作中,我们调查长尾图像识别SB,发现尾类在SB中遭受更严重的伤害,从而损害此类代表性不足的类别的总体性表现。我们从经验上报告,自我监督学习(SSL)可以减少SB,并通过丰富从尾巴样品提取的特征,从而更好地利用这类稀有样本来补充受监督的对应方。然而,标准SSL方法的设计没有明确考虑以类别为单位的内在数据分布,可能不是长尾部分布数据的最佳方法。为了解决这一局限性,我们提出了一种针对数据不平衡的新型SSL方法。它通过三重不同层次,即整体、部分和扩大层次,利用SSL来增强预测性复杂模式的学习,从而有可能克服严重尾部数据的SB。在五大尾部基准数据集方面的定量和定性实验结果都能够有效地减少我们的方法。