In defense-related remote sensing applications, such as vehicle detection on satellite imagery, supervised learning requires a huge number of labeled examples to reach operational performances. Such data are challenging to obtain as it requires military experts, and some observables are intrinsically rare. This limited labeling capability, as well as the large number of unlabeled images available due to the growing number of sensors, make object detection on remote sensing imagery highly relevant for self-supervised learning. We study in-domain self-supervised representation learning for object detection on very high resolution optical satellite imagery, that is yet poorly explored. For the first time to our knowledge, we study the problem of label efficiency on this task. We use the large land use classification dataset Functional Map of the World to pretrain representations with an extension of the Momentum Contrast framework. We then investigate this model's transferability on a real-world task of fine-grained vehicle detection and classification on Preligens proprietary data, which is designed to be representative of an operational use case of strategic site surveillance. We show that our in-domain self-supervised learning model is competitive with ImageNet pretraining, and outperforms it in the low-label regime.
翻译:在与国防有关的遥感应用中,如在卫星图像上探测车辆,有监督的学习需要大量贴标签的例子,才能达到操作性能。这些数据在获得需要军事专家时具有挑战性,而且有些观测数据也基本很少。这种有限的标签能力,以及由于传感器数量不断增加而可获得的大量无标签图像,使得遥感图像中的物体探测对自我监督学习具有高度相关性。我们研究在非常高分辨率光学卫星图像上进行物体探测的自我监督的代表学习,这种研究还没有得到很好地探讨。我们第一次了解,我们研究了这一任务中的标签效率问题。我们利用大型土地使用分类数据集世界功能图来进行预演,扩展了移动式相抗力框架。然后我们调查这一模型在精细的车辆探测和对普列根斯专利数据进行分类的实际任务中的可转移性,该模型旨在代表战略网站监测的实际使用案例。我们通过内部自我监督的自我监督模型在图像网络前培训中具有竞争力,并显示其低模版的测试系统具有竞争力。