Semi-supervised learning (SSL) addresses the lack of labeled data by exploiting large unlabeled data through pseudolabeling. However, in the extremely low-label regime, pseudo labels could be incorrect, a.k.a. the confirmation bias, and the pseudo labels will in turn harm the network training. Recent studies combined finetuning (FT) from pretrained weights with SSL to mitigate the challenges and claimed superior results in the low-label regime. In this work, we first show that the better pretrained weights brought in by FT account for the state-of-the-art performance, and importantly that they are universally helpful to off-the-shelf semi-supervised learners. We further argue that direct finetuning from pretrained weights is suboptimal due to covariate shift and propose a contrastive target pretraining step to adapt model weights towards target dataset. We carried out extensive experiments on both classification and segmentation tasks by doing target pretraining then followed by semi-supervised finetuning. The promising results validate the efficacy of target pretraining for SSL, in particular in the low-label regime.
翻译:半监督学习(SSL) 解决标签数据缺乏的问题,办法是通过假贴标签来利用大型未贴标签数据。然而,在极低的标签制度下,假标签可能是不正确的,a.k.a.确认偏差和假标签反过来又会损害网络培训。最近的一些研究将预先培训的重量与SSL合并,以缓解挑战,并声称低标签制度中的优异结果。在这项工作中,我们首先显示,FT账户为最新业绩带来的更好的预先培训重量,重要的是,它们对于现成的半监督学习者普遍有帮助。我们进一步认为,对预培训重量的直接微调是次优的,因为相互变换,并提出一个对比性的目标培训前步骤,使模型重量适应目标数据集。我们进行了广泛的分类和分化任务实验,先进行目标培训,然后进行半监督的微调。有希望的结果验证了SSL(特别是低标签制度)目标培训前训练的效果。