Contrastive representation learning has proven to be an effective self-supervised learning method. Most successful approaches are based on the Noise Contrastive Estimation (NCE) paradigm and consider different views of an instance as positives and other instances as noise that positives should be contrasted with. However, all instances in a dataset are drawn from the same distribution and share underlying semantic information that should not be considered as noise. We argue that a good data representation contains the relations, or semantic similarity, between the instances. Contrastive learning implicitly learns relations but considers the negatives as noise which is harmful to the quality of the learned relations and therefore the quality of the representation. To circumvent this issue we propose a novel formulation of contrastive learning using semantic similarity between instances called Similarity Contrastive Estimation (SCE). Our training objective can be considered as soft contrastive learning. Instead of hard classifying positives and negatives, we propose a continuous distribution to push or pull instances based on their semantic similarities. The target similarity distribution is computed from weak augmented instances and sharpened to eliminate irrelevant relations. Each weak augmented instance is paired with a strong augmented instance that contrasts its positive while maintaining the target similarity distribution. Experimental results show that our proposed SCE outperforms its baselines MoCov2 and ReSSL on various datasets and is competitive with state-of-the-art algorithms on the ImageNet linear evaluation protocol.
翻译:对比性代表制学习已证明是一种有效的自我监督学习方法。 多数成功的方法都基于“ 噪音对比性估计( NCE) ” 模式,并将对一个实例的不同观点视为积极性,而将其他实例视为积极性应与之对比的噪音。 但是,数据集中的所有实例都来自同样的分布,并分享不应被视为噪音的语义信息。 我们争辩说, 良好的数据代表制包含不同实例之间的关系或语义相似性。 对比性学习隐含着学习关系,但认为负面的噪音有害于所学关系的质量,因此也认为代表性的质量。 为绕过这一问题,我们建议使用被称为相似性相对性刺激(SCE)的语义相似性来制定新的对比性学习模式。 我们的培训目标可以被视为软化的对比性学习。 我们建议, 将正性和负性进行硬性分类, 依据其语义相似性相似性来持续发布或拉动性实例。 目标相似性分布来自较弱的递增性实例, 并且更锐化到消除不相近性的图像。 每个实验性实例都显示SCE的对比性。