We are interested in representation learning in self-supervised, supervised, or semi-supervised settings. The prior work on applying mean-shift idea for self-supervised learning, MSF, generalizes the BYOL idea by pulling a query image to not only be closer to its other augmentation, but also to the nearest neighbors (NNs) of its other augmentation. We believe the learning can benefit from choosing far away neighbors that are still semantically related to the query. Hence, we propose to generalize MSF algorithm by constraining the search space for nearest neighbors. We show that our method outperforms MSF in SSL setting when the constraint utilizes a different augmentation of an image, and outperforms PAWS in semi-supervised setting with less training resources when the constraint ensures the NNs have the same pseudo-label as the query.
翻译:我们有兴趣在自我监督、监督或半监督的环境下进行代表性学习。 先前关于应用自监督学习的中变式想法的工作,MSF, 将BYOL的想法概括化, 不仅拉近查询图像, 更接近其其他增强, 也拉近其其他增强的近邻。 我们认为, 选择远方邻居, 而这些邻居仍然与查询有内在联系, 有助于学习学习。 因此, 我们提议通过限制近邻的搜索空间, 推广MSF 算法。 我们显示, 当限制使用不同放大图像时, 我们的方法优于SSL 设置的MSF, 在半监督的环境下, 当限制确保无核武器国家拥有与查询相同的假标签时, 我们的方法优于 MSF 。