Semi-Supervised Object Detection (SSOD), aiming to explore unlabeled data for boosting object detectors, has become an active task in recent years. However, existing SSOD approaches mainly focus on horizontal objects, leaving multi-oriented objects that are common in aerial images unexplored. This paper proposes a novel Semi-supervised Oriented Object Detection model, termed SOOD, built upon the mainstream pseudo-labeling framework. Towards oriented objects in aerial scenes, we design two loss functions to provide better supervision. Focusing on the orientations of objects, the first loss regularizes the consistency between each pseudo-label-prediction pair (includes a prediction and its corresponding pseudo label) with adaptive weights based on their orientation gap. Focusing on the layout of an image, the second loss regularizes the similarity and explicitly builds the many-to-many relation between the sets of pseudo-labels and predictions. Such a global consistency constraint can further boost semi-supervised learning. Our experiments show that when trained with the two proposed losses, SOOD surpasses the state-of-the-art SSOD methods under various settings on the DOTA-v1.5 benchmark. The code will be available at https://github.com/HamPerdredes/SOOD.
翻译:半监督目标检测(SSOD)旨在利用未标注数据来提高目标检测器的性能,在最近的研究中已成为活跃的领域。然而,现有的半监督目标检测方法主要集中在水平对象上,在航空图像中普遍存在的多向对象则鲜有探讨。本文提出了一种新的半监督定向目标检测模型,称为SOOD,它建立在主流的伪标签框架之上。针对航空场景中的定向对象,我们设计了两个损失函数来提供更好的监督。第一个损失函数侧重于对象的方向,通过自适应权重来规范每个伪标签-预测对之间的一致性,其中权重基于它们的方向差。第二个损失函数则侧重于图像的布局,规范相似性,明确构建伪标签集和预测集之间的多对多关系。这种全局一致性约束进一步提高了半监督学习的性能。我们的实验表明,在使用了两个提出的损失训练后,SOOD在DOTA-v1.5数据集上各种设置下都超过了最先进的SSOD方法。代码将在https://github.com/HamPerdredes/SOOD上公开发布。