Contrastive learning (CL) is one of the most successful paradigms for self-supervised learning (SSL). In a principled way, it considers two augmented ``views'' of the same image as positive to be pulled closer, and all other images negative to be pushed further apart. However, behind the impressive success of CL-based techniques, their formulation often relies on heavy-computation settings, including large sample batches, extensive training epochs, etc. We are thus motivated to tackle these issues and aim at establishing a simple, efficient, and yet competitive baseline of contrastive learning. Specifically, we identify, from theoretical and empirical studies, a noticeable negative-positive-coupling (NPC) effect in the widely used cross-entropy (InfoNCE) loss, leading to unsuitable learning efficiency with respect to the batch size. Indeed the phenomenon tends to be neglected in that optimizing infoNCE loss with a small-size batch is effective in solving easier SSL tasks. By properly addressing the NPC effect, we reach a decoupled contrastive learning (DCL) objective function, significantly improving SSL efficiency. DCL can achieve competitive performance, requiring neither large batches in SimCLR, momentum encoding in MoCo, or large epochs. We demonstrate the usefulness of DCL in various benchmarks, while manifesting its robustness being much less sensitive to suboptimal hyperparameters. Notably, our approach achieves $66.9\%$ ImageNet top-1 accuracy using batch size 256 within 200 epochs pre-training, outperforming its baseline SimCLR by $5.1\%$. With further optimized hyperparameters, DCL can improve the accuracy to $68.2\%$. We believe DCL provides a valuable baseline for future contrastive learning-based SSL studies.
翻译:对比学习(CL)是自我监督学习(SSL)最成功的范例之一。 以原则性的方式,它认为两种强化的“ 视图” 图像与正相近的相同图像将拉近,而所有其他图像则将进一步推开。 但是,在基于CL的技术取得令人印象深刻的成功之后,它们的配方往往依赖于重算设置,包括大量样本批量、广泛的培训时代等等。 因此,我们有动力解决这些问题,目的是建立一个简单、高效和有竞争力的对比学习基线。 具体地说,我们从理论和经验研究中,发现两种“视图”的图像与正比的图像相近的相同,而“观点”的相同,在广泛使用的交叉opropy(InFONCE)损失中,明显的负正比(NPC)效应明显负比(NPC)效应,在SLDR(SL)下,在高的SL(SL)下,在SL(SL)下,在大幅的SL(SL)下,在SL(SL)下,在SD(SL)下,在高的SL(SL)下)下,可以实现高的竞争力。