This paper proposes a novel contrastive learning framework, coined as Self-Contrastive (SelfCon) Learning, that self-contrasts within multiple outputs from the different levels of a network. We confirmed that SelfCon loss guarantees the lower bound of mutual information (MI) between the intermediate and last representations. Besides, we empirically showed, via various MI estimators, that SelfCon loss highly correlates to the increase of MI and better classification performance. In our experiments, SelfCon surpasses supervised contrastive (SupCon) learning without the need for a multi-viewed batch and with the cheaper computational cost. Especially on ResNet-18, we achieved top-1 classification accuracy of 76.45% for the CIFAR-100 dataset, which is 2.87% and 4.36% higher than SupCon and cross-entropy loss, respectively. We found that mitigating both vanishing gradient and overfitting issue makes our method outperform the counterparts.
翻译:本文提出了一个新颖的对比式学习框架,即自相残杀(自相残杀(自相残杀)学习(自相残杀),在网络不同层次的多种产出中自相残杀(自相残杀),我们确认,自相残杀(自相残杀)能保证中间和最后代表之间相互信息的较低范围。此外,我们通过各种自相残杀的估测员,从经验上表明,自相残杀(自相残杀)与MI的增加和更好的分类性能密切相关。在我们的实验中,自相残杀(自相残杀(自相残杀(自相残杀(自相残杀))超过受监督的对比性学习(自相残杀(自相残杀),不需要多目批和廉价的计算费用。 特别是在ResNet-18上,我们为CIFAR-100数据集实现了76.45%的上一级分类精确度,分别比Supn和跨编损失高出2.87%和4.36%。我们发现,减缓梯度问题使我们的方法超越了对应者。