One paradigm for learning from few labeled examples while making best use of a large amount of unlabeled data is unsupervised pretraining followed by supervised fine-tuning. Although this paradigm uses unlabeled data in a task-agnostic way, in contrast to common approaches to semi-supervised learning for computer vision, we show that it is surprisingly effective for semi-supervised learning on ImageNet. A key ingredient of our approach is the use of big (deep and wide) networks during pretraining and fine-tuning. We find that, the fewer the labels, the more this approach (task-agnostic use of unlabeled data) benefits from a bigger network. After fine-tuning, the big network can be further improved and distilled into a much smaller one with little loss in classification accuracy by using the unlabeled examples for a second time, but in a task-specific way. The proposed semi-supervised learning algorithm can be summarized in three steps: unsupervised pretraining of a big ResNet model using SimCLRv2, supervised fine-tuning on a few labeled examples, and distillation with unlabeled examples for refining and transferring the task-specific knowledge. This procedure achieves 73.9% ImageNet top-1 accuracy with just 1% of the labels ($\le$13 labeled images per class) using ResNet-50, a $10\times$ improvement in label efficiency over the previous state-of-the-art. With 10% of labels, ResNet-50 trained with our method achieves 77.5% top-1 accuracy, outperforming standard supervised training with all of the labels.
翻译:从几个贴标签的示例中学习的范例之一,在最佳使用大量未贴标签的数据的同时,在最佳使用大量未贴标签的数据时,没有监督的预培训,随后进行监管的微调。虽然这个范例使用未贴标签的数据的方式是任务不可知的,但与半监视的计算机视觉学习的通用方法相反,我们显示,它对于半监视的图像网学习是惊人的。我们的方法的一个关键要素是在预培训和微调整期间使用大型(深度和广度)网络网络。我们发现,标签越少,这个方法(无标签的50数据的保密使用)就越能从一个更大的网络中受益。在微调后,大网络可以进一步改进,并升至一个小得多,而分类精准度则在第二次使用未贴标签的示例时,但以特定任务的方式。 拟议的半监控学习算法可以分为三个步骤:使用SimCLRv2对大型 ResNet 模型进行未经校准的预培训,对一些标签的50美元准确性数据进行监管的精确性调整,在1个特定标签的示例中,用已经过培训的10 % 升级的升级的升级后,在1个标签上,用1个标签的升级的升级的升级的升级的升级的升级的升级的升级的高级程序,在1个升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的升级的流程。