Semi-supervised learning techniques are gaining popularity due to their capability of building models that are effective, even when scarce amounts of labeled data are available. In this paper, we present a framework and specific tasks for self-supervised pretraining of \textit{multichannel} models, such as the fusion of multispectral and synthetic aperture radar images. We show that the proposed self-supervised approach is highly effective at learning features that correlate with the labels for land cover classification. This is enabled by an explicit design of pretraining tasks which promotes bridging the gaps between sensing modalities and exploiting the spectral characteristics of the input. In a semi-supervised setting, when limited labels are available, using the proposed self-supervised pretraining, followed by supervised finetuning for land cover classification with SAR and multispectral data, outperforms conventional approaches such as purely supervised learning, initialization from training on ImageNet and other recent self-supervised approaches.
翻译:半受监督的学习技术由于建立有效模型的能力而越来越受欢迎,即使有很少的标签数据,在本文中,我们提出了一个框架和具体任务,用于对\ textit{多通道]模型进行自我监督的预培训,例如将多光谱和合成孔径雷达图像聚合起来。我们表明,拟议的自监督方法对于与土地覆盖分类标签相关的学习特点非常有效。这得益于明确设计培训前任务,这种任务有助于弥合遥感模式与利用输入的光谱特征之间的差距。在半监督环境下,如果有有限的标签,则使用拟议的自监督前培训,随后对土地覆盖分类与合成孔径雷达和多光谱数据进行监管的微调,比常规方法(如纯监督学习、从关于图像网络的培训和其他最近自我监督的方法中初始化)更完善。