As a self-supervised learning paradigm, contrastive learning has been widely used to pre-train a powerful encoder as an effective feature extractor for various downstream tasks. This process requires numerous unlabeled training data and computational resources, which makes the pre-trained encoder become valuable intellectual property of the owner. However, the lack of a priori knowledge of downstream tasks makes it non-trivial to protect the intellectual property of the pre-trained encoder by applying conventional watermarking methods. To deal with this problem, in this paper, we introduce AWEncoder, an adversarial method for watermarking the pre-trained encoder in contrastive learning. First, as an adversarial perturbation, the watermark is generated by enforcing the training samples to be marked to deviate respective location and surround a randomly selected key image in the embedding space. Then, the watermark is embedded into the pre-trained encoder by further optimizing a joint loss function. As a result, the watermarked encoder not only performs very well for downstream tasks, but also enables us to verify its ownership by analyzing the discrepancy of output provided using the encoder as the backbone under both white-box and black-box conditions. Extensive experiments demonstrate that the proposed work enjoys pretty good effectiveness and robustness on different contrastive learning algorithms and downstream tasks, which has verified the superiority and applicability of the proposed work.
翻译:作为自我监督的学习范式,对比式学习被广泛用于培训强大的编码器,作为各种下游任务的有效特征提取器。这一过程需要无数未贴标签的培训数据和计算资源,使经过事先训练的编码器成为所有者的宝贵知识产权。然而,由于缺乏先验的下游任务知识,因此,通过应用传统的水标记方法来保护经过训练的编码器的知识产权是非边际的。为了处理这一问题,我们在本文件中引入了AWEncoder,这是在对比性学习中为事先训练过的编码师打水标记的一种对抗性方法。首先,作为对抗性渗透,水标记是通过执行培训样本来生成的,以便标记不同的地点,并围绕嵌入空间随机选择的关键图像。随后,水标记通过进一步优化联合损失功能而嵌入经过训练前的编码器。因此,水标记编码器不仅在下游任务中很好地运行,而且还使我们能够通过在拟议的基础测试中分析产出的深度差异,从而核实其所有权,从而在拟议的基础测试中展示了拟议中,从而展示了对成果的深度测试。