Contrastive learning has emerged as an essential approach for self-supervised learning in computer vision. The central objective of contrastive learning is to maximize the similarities between two augmented versions of the same image (positive pairs), while minimizing the similarities between different images (negative pairs). Recent studies have demonstrated that harder negative samples, i.e., those that are difficult to distinguish from anchor sample, play a more critical role in contrastive learning. In this paper, we propose a novel featurelevel method, namely sampling synthetic hard negative samples for contrastive learning (SSCL), to exploit harder negative samples more effectively. Specifically, 1) we generate more and harder negative samples by mixing negative samples, and then sample them by controlling the contrast of anchor sample with the other negative samples. 2) Considering that the negative samples obtained by sampling may have the problem of false negative samples, we further debias the negative samples. Our proposed method improves the classification performance on different image datasets and can be readily applied to existing methods.
翻译:对比学习已经成为计算机视觉中自监督学习的重要方法,其中核心目标是最大化同一图像的两个增强版本之间的相似性(正对),同时最小化不同图像之间的相似性(负对)。近期的研究已经证明,更难的负样本,即难以与锚样本区分的样本,在对比学习中起着更加关键的作用。在本文中,我们提出了一种新的特征级方法,即利用合成难负样本用于对比学习(SSCL),更加有效地利用更难的负样本。具体而言,1)我们通过混合负样本来生成更多和更难的负样本,然后通过控制锚样本与其他负样本之间的对比度来对它们进行采样。2)考虑到通过采样获得的负样本可能存在误差,我们进一步纠正了负样本。我们提出的方法提高了不同图像数据集上的分类性能,并可轻松应用于现有方法中。