Recently, self-supervised learning has attracted attention due to its remarkable ability to acquire meaningful representations for classification tasks without using semantic labels. This paper introduces a self-supervised learning framework suitable for hyperspectral images that are inherently challenging to annotate. The proposed framework architecture leverages cross-domain CNN, allowing for learning representations from different hyperspectral images with varying spectral characteristics and no pixel-level annotation. In the framework, cross-domain representations are learned via contrastive learning where neighboring spectral vectors in the same image are clustered together in a common representation space encompassing multiple hyperspectral images. In contrast, spectral vectors in different hyperspectral images are separated into distinct clusters in the space. To verify that the learned representation through contrastive learning is effectively transferred into a downstream task, we perform a classification task on hyperspectral images. The experimental results demonstrate the advantage of the proposed self-supervised representation over models trained from scratch or other transfer learning methods.
翻译:最近,自我监督的学习因其在不使用语义标签的情况下获得对分类任务有意义的有意义表述的非凡能力而引起人们的注意。本文件介绍了一个适合超光谱图像的自监督学习框架,这种超光谱图像对笔记本具有固有的挑战性。拟议框架架构利用了跨域CNN, 从而可以从具有不同频谱特征和没有像素级别的像素级注释的不同超光谱图像中学习演示。在这个框架中,通过对比学习学习,学习了跨域演示,将同一图像中的相邻光谱矢量聚集在一个包含多个超光谱图像的共同代表空间。相比之下,不同超光谱图像中的光谱矢量被分离成不同的空间集群。为了核实通过对比学习而获得的演示被有效地转移到下游任务,我们开展了超光谱图像的分类任务。实验结果表明,拟议的自监督演示对于从抓或其他传输学习方法中训练的模型的优势。