Unsupervised learning has always been appealing to machine learning researchers and practitioners, allowing them to avoid an expensive and complicated process of labeling the data. However, unsupervised learning of complex data is challenging, and even the best approaches show much weaker performance than their supervised counterparts. Self-supervised deep learning has become a strong instrument for representation learning in computer vision. However, those methods have not been evaluated in a fully unsupervised setting. In this paper, we propose a simple scheme for unsupervised classification based on self-supervised representations. We evaluate the proposed approach with several recent self-supervised methods showing that it achieves competitive results for ImageNet classification (39% accuracy on ImageNet with 1000 clusters and 46% with overclustering). We suggest adding the unsupervised evaluation to a set of standard benchmarks for self-supervised learning. The code is available at https://github.com/Randl/kmeans_selfsuper
翻译:无监督的学习一直吸引着机器学习研究人员和从业者,使他们可以避免一个昂贵和复杂的数据标签程序。然而,未经监督的复杂数据的学习具有挑战性,甚至最佳方法也显示其业绩比受监督的同行差得多。自我监督的深层次学习已成为计算机愿景中体现学习的有力工具。然而,这些方法还没有在完全不受监督的环境中进行评估。在本文件中,我们提出了一个基于自我监督的表述方式进行不受监督分类的简单计划。我们用最近一些自我监督的方法评估了拟议方法,显示其在图像网络分类上取得了竞争性结果(在图像网络上,39%的精确度为1000个集群,46%的精确度为超集群)。我们建议将未经监督的评价添加到一套自我监督学习的标准基准中。该代码可在 https://github.com/Randl/kpoles_selfeurvey查阅。