Contrastive representation learning has proven to be an effective self-supervised learning method. Most successful approaches are based on Noise Contrastive Estimation (NCE) and use different views of an instance as positives that should be contrasted with other instances, called negatives, that are considered as noise. However, several instances in a dataset are drawn from the same distribution and share underlying semantic information. A good data representation should contain relations, or semantic similarity, between the instances. Contrastive learning implicitly learns relations but considering all negatives as noise harms the quality of the learned relations. To circumvent this issue, we propose a novel formulation of contrastive learning using semantic similarity between instances called Similarity Contrastive Estimation (SCE). Our training objective is a soft contrastive learning one. Instead of hard classifying positives and negatives, we estimate from one view of a batch a continuous distribution to push or pull instances based on their semantic similarities. This target similarity distribution is sharpened to eliminate noisy relations. The model predicts for each instance, from another view, the target distribution while contrasting its positive with negatives. Experimental results show that SCE is Top-1 on the ImageNet linear evaluation protocol at 100 pretraining epochs with 72.1% accuracy and is competitive with state-of-the-art algorithms by reaching 75.4% for 200 epochs with multi-crop. We also show that SCE is able to generalize to several tasks. Source code is available here: https://github.com/CEA-LIST/SCE.
翻译:有效的自我监督学习方法。 大多数成功的方法都基于噪音对比估计( NCE), 并且使用不同的观点作为正数, 与其他的例子( 称为负数, 被认为是噪音) 相对照。 然而, 数据集中的一些实例来自相同的分布, 并分享基本的语义信息。 良好的数据代表应该包含不同实例之间的关系或语义相似性。 相近学习暗中学习关系, 但将所有负数都视为噪音会损害所学关系的质量。 为了绕过这个问题, 我们建议采用新颖的对比性学习模式, 使用类似词( 类似对比对比动画( SCE) 。 我们的培训目标是软化的对比性学习 。 我们从一种基于语义相似性的分批量连续分布到推动或拉动事件。 这个目标相似性分布更加精锐化, 能够消除噪音关系。 每个例子的模型预测, 从另一个角度, 目标分布比重( SCEEO) 与 AS- AS- AS- AS- AS- AS- AS- imalal imalal 对比性结果。