Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets. However, large-scale annotations are time-consuming and labor-exhaustive to obtain on realistic tasks. To mitigate the requirement for labeled data, self-training is widely used in both academia and industry by pseudo labeling on readily-available unlabeled data. Despite its popularity, pseudo labeling is well-believed to be unreliable and often leads to training instability. Our experimental studies further reveal that the performance of self-training is biased due to data sampling, pre-trained models, and training strategies, especially the inappropriate utilization of pseudo labels. To this end, we propose Debiased, in which the generation and utilization of pseudo labels are decoupled by two independent heads. To further improve the quality of pseudo labels, we introduce a worst-case estimation of pseudo labeling and seamlessly optimize the representations to avoid the worst-case. Extensive experiments justify that the proposed Debiased not only yields an average improvement of $14.4$\% against state-of-the-art algorithms on $11$ tasks (covering generic object recognition, fine-grained object recognition, texture classification, and scene classification) but also helps stabilize training and balance performance across classes.
翻译:深心神经网络在大型标签数据集的帮助下,在一系列广泛的任务上取得了显著的成绩。然而,大型说明耗时费时,而且劳动繁忙,无法完成现实的任务。为了减轻对标签数据的要求,学术界和行业都广泛采用自我培训,在现成的未贴标签数据上打假标签。尽管伪标签很受欢迎,但人们认为其不可靠,往往导致培训不稳定。我们的实验研究进一步表明,自我培训的绩效不仅由于数据抽样、预先培训的模式和培训战略,特别是假标签的不当使用而有偏差。为此,我们提议Debiased,其中伪标签的生成和使用被两个独立头目拆分。为了进一步提高伪标签的质量,我们引入了最坏的假标签估计,并且无缝地优化了表述以避免最坏的情况。广泛的实验证明,拟议的贬损不仅因为数据抽样、预先培训模型和培训战略,特别是假标签的使用不当。为此,我们提议Debiased,其中伪标签的生成和使用被两个独立头分解。为了进一步提高伪标签的质量,我们引入了最坏的假设性估计,并且完美地优化地优化地优化地优化了模拟分类,从而避免出现最坏的情况。