Unsupervised visual representation learning has gained much attention from the computer vision community because of the recent achievement of contrastive learning. Most of the existing contrastive learning frameworks adopt the instance discrimination as the pretext task, which treating every single instance as a different class. However, such method will inevitably cause class collision problems, which hurts the quality of the learned representation. Motivated by this observation, we introduced a weakly supervised contrastive learning framework (WCL) to tackle this issue. Specifically, our proposed framework is based on two projection heads, one of which will perform the regular instance discrimination task. The other head will use a graph-based method to explore similar samples and generate a weak label, then perform a supervised contrastive learning task based on the weak label to pull the similar images closer. We further introduced a K-Nearest Neighbor based multi-crop strategy to expand the number of positive samples. Extensive experimental results demonstrate WCL improves the quality of self-supervised representations across different datasets. Notably, we get a new state-of-the-art result for semi-supervised learning. With only 1\% and 10\% labeled examples, WCL achieves 65\% and 72\% ImageNet Top-1 Accuracy using ResNet50, which is even higher than SimCLRv2 with ResNet101.
翻译:未经监督的视觉代表学习由于最近的对比性学习成果而引起了计算机视觉视觉视觉界的极大关注。 大部分现有的对比性学习框架将实例歧视作为借口任务, 将每个实例视为不同的类别。 然而, 这种方法将不可避免地造成阶级碰撞问题, 从而伤害了所学代表的质量。 我们受此观察的驱动, 引入了一个监督不力的对比性学习框架( WWL) 来解决这一问题。 具体地说, 我们的拟议框架基于两个投影头, 其中之一将执行常规实例歧视任务。 另一个首级将使用基于图表的方法来探索相似的样本并生成一个薄弱的标签, 然后在薄弱标签的基础上执行监督的对比性学习任务, 来拉近相似的图像。 我们还引入了一个基于K- Nearest Neighbor 的多作物战略, 以扩大正面样本的数量。 广泛的实验结果显示WCLF在不同数据集中提高了自我监督的表达质量。 值得注意的是, 我们将获得一个新的状态艺术结果, 用于半超导性学习。 以1+++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++