Training object detection models usually requires instance-level annotations, such as the positions and labels of all objects present in each image. Such supervision is unfortunately not always available and, more often, only image-level information is provided, also known as weak supervision. Recent works have addressed this limitation by leveraging knowledge from a richly annotated domain. However, the scope of weak supervision supported by these approaches has been very restrictive, preventing them to use all available information. In this work, we propose ProbKT, a framework based on probabilistic logical reasoning that allows to train object detection models with arbitrary types of weak supervision. We empirically show on different datasets that using all available information is beneficial as our ProbKT leads to significant improvement on target domain and better generalization compared to existing baselines. We also showcase the ability of our approach to handle complex logic statements as supervision signal.
翻译:培训对象探测模型通常需要实例级说明,例如每个图像中所有对象的位置和标签。这种监督不幸并不总是可以获得,而且更经常地只提供图像级信息,也称为薄弱监督。最近的一些工作通过利用一个备注丰富的领域知识解决了这一局限性。然而,这些方法所支持的薄弱监督范围非常狭窄,无法使用所有可用信息。在这项工作中,我们提议ProbKT,这是一个基于概率逻辑逻辑推理的框架,能够以任意类型的薄弱监督来培训对象检测模型。我们的经验显示,使用所有可用信息都是有益的,因为我们的ProbKT在目标领域取得了显著改进,并且比现有的基线更加普遍化。我们还展示了我们处理复杂逻辑说明作为监督信号的方法的能力。</s>