We propose a non-iterative method to optimize pseudo-labeling thresholds for learning object detection from a collection of low-cost datasets, each of which is annotated for only a subset of all the object classes. A popular approach to this problem is first to train teacher models and then to use their confident predictions as pseudo ground-truth labels when training a student model. To obtain the best result, however, thresholds for prediction confidence must be adjusted. This process typically involves iterative search and repeated training of student models and is time-consuming. Therefore, we develop a method to optimize the thresholds without iterative optimization by maximizing the $F_\beta$-score on a validation dataset, which measures the quality of pseudo labels and can be measured without training a student model. We experimentally demonstrate that our proposed method achieves an mAP comparable to that of grid search on the COCO and VOC datasets.
翻译:我们建议采用非象征性方法,从低成本数据集集中优化学习对象探测的假标签阈值,每个数据集仅说明所有对象类的一个子集。这个问题的流行做法是首先培训教师模型,然后在培训学生模型时将其自信预测作为假地面真实标签。然而,为了取得最佳结果,必须调整预测信心阈值。这一过程通常涉及反复搜索和重复培训学生模型,而且耗时。因此,我们开发了一种方法,在不重复优化的情况下优化阈值,在验证数据集中最大限度地使用美元-贝塔元核心,以衡量假标签的质量,并在不培训学生模型的情况下进行测量。我们实验性地证明,我们提出的方法达到了与COCO和VOC数据集网格搜索相仿的 mAP。