To improve instance-level detection/segmentation performance, existing self-supervised and semi-supervised methods extract either task-unrelated or task-specific training signals from unlabeled data. We show that these two approaches, at the two extreme ends of the task-specificity spectrum, are suboptimal for the task performance. Utilizing too little task-specific training signals causes underfitting to the ground-truth labels of downstream tasks, while the opposite causes overfitting to the ground-truth labels. To this end, we propose a novel Class-Agnostic Semi-Supervised Learning (CA-SSL) framework to achieve a more favorable task-specificity balance in extracting training signals from unlabeled data. CA-SSL has three training stages that act on either ground-truth labels (labeled data) or pseudo labels (unlabeled data). This decoupling strategy avoids the complicated scheme in traditional SSL methods that balances the contributions from both data types. Especially, we introduce a warmup training stage to achieve a more optimal balance in task specificity by ignoring class information in the pseudo labels, while preserving localization training signals. As a result, our warmup model can better avoid underfitting/overfitting when fine-tuned on the ground-truth labels in detection and segmentation tasks. Using 3.6M unlabeled data, we achieve a significant performance gain of 4.7% over ImageNet-pretrained baseline on FCOS object detection. In addition, our warmup model demonstrates excellent transferability to other detection and segmentation frameworks.
翻译:为改善试级检测/分类性能,现有的自我监管和半监督的网络性能,从未贴标签的数据中提取与任务无关或特定任务的培训信号。 我们显示,这两种方法,在任务特定频谱的两个极端端,对于任务性能来说是不最优化的。 利用过少的任务特定培训信号导致与下游任务的地面真实标签不匹配,而相反的原因则与地面真实性标签不符。 为此,我们提议了一个新型的类级 Agnoti Sid- Supervisial Learning (CA-SSL) 框架,以便在从未贴标签的数据中提取培训信号时实现更有利的任务特定性平衡。 CA-SSL有三个培训阶段在地面真实性标签(标签数据)或假贴标签(未贴标签数据)上运作。 这种分解战略避免了传统 SSLL 方法的复杂机制,平衡了我们两种数据类型的贡献。 特别是,我们引入了一个更暖化的培训阶段, 以在任务性稳定性评估性框架中实现更优化的平衡, 忽略了实地检测结果,同时在类中, 保持了更精确的标签。