Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets. Yet these datasets are time-consuming and labor-exhaustive to obtain on realistic tasks. To mitigate the requirement for labeled data, self-training is widely used in semi-supervised learning by iteratively assigning pseudo labels to unlabeled samples. Despite its popularity, self-training is well-believed to be unreliable and often leads to training instability. Our experimental studies further reveal that the bias in semi-supervised learning arises from both the problem itself and the inappropriate training with potentially incorrect pseudo labels, which accumulates the error in the iterative self-training process. To reduce the above bias, we propose Debiased Self-Training (DST). First, the generation and utilization of pseudo labels are decoupled by two parameter-independent classifier heads to avoid direct error accumulation. Second, we estimate the worst case of self-training bias, where the pseudo labeling function is accurate on labeled samples, yet makes as many mistakes as possible on unlabeled samples. We then adversarially optimize the representations to improve the quality of pseudo labels by avoiding the worst case. Extensive experiments justify that DST achieves an average improvement of 6.3% against state-of-the-art methods on standard semi-supervised learning benchmark datasets and 18.9%$ against FixMatch on 13 diverse tasks. Furthermore, DST can be seamlessly adapted to other self-training methods and help stabilize their training and balance performance across classes in both cases of training from scratch and finetuning from pre-trained models.
翻译:深心神经网络在大型标签数据集的帮助下,在一系列广泛的任务上取得了显著的成绩。然而,这些数据集耗时费力,而且为了获得现实的任务,这些数据集耗费大量人力,因此难以获得现实的任务。为了减轻标签数据的要求,自我培训被广泛用于半监督的学习中,通过迭代将假标签划为无标签的样本进行。尽管自我培训受到欢迎,但人们相信它不可靠,往往导致培训不稳定。我们的实验研究进一步显示,半监督学习的偏差来自问题本身和可能错误的假标签的不适当培训,这在迭代自我培训过程中积累了错误。为了减少上述偏差,我们提议降低标签数据的偏差自我培训(DST)。首先,将假标签的生成和使用由两个依赖参数的分类头进行拆分解,以避免直接的错误积累。第二,我们估计自培训偏差的最坏的例子,即标签的多样化标签功能准确无误,但在未标定的样本中尽可能有许多错误,但避免了不正确的假假标签的假标签。然后,我们提议,在最差的自我培训前的自我训练方法方面,在最差的模型上优化的模型上,在最差的模型上,用最差的模型上,可以改进。