Graph Neural Networks (GNNs) have achieved promising performance in various real-world applications. Building a powerful GNN model is not a trivial task, as it requires a large amount of training data, powerful computing resources, and human expertise in fine-tuning the model. What is more, with the development of adversarial attacks, e.g., model stealing attacks, GNNs raise challenges to model authentication. To avoid copyright infringement on GNNs, it is necessary to verify the ownership of the GNN models. In this paper, we present a watermarking framework for GNNs for both graph and node classification tasks. We 1) design two strategies to generate watermarked data for the graph classification task and one for the node classification task, 2) embed the watermark into the host model through training to obtain the watermarked GNN model, and 3) verify the ownership of the suspicious model in a black-box setting. The experiments show that our framework can verify the ownership of GNN models with a very high probability (around $95\%$) for both tasks. Finally, we experimentally show that our watermarking approach is robust against two model modifications and an input reformation defense against backdoor attacks.
翻译:神经网络图(GNN)在各种现实应用中取得了有希望的绩效。 建立一个强大的GNN模型并不是一项微不足道的任务,因为它需要大量的培训数据、强大的计算资源和精通人员对模型进行微调。 此外,随着对抗性攻击的发展,例如模型盗窃攻击,GNNS对模型认证提出了挑战。为了避免对GNN的版权侵犯,有必要核实GNN模型的所有权。在本文中,我们为图形和节点分类任务提出了一个全球NNN模型的水标记框架。我们1)设计了两种战略,为图形分类任务和节点分类任务生成水标记数据;2)通过培训将水标记纳入主机模型,以获得水标记GNN模型模型模型模型,3)在黑箱设置中核查可疑模型的所有权。实验表明,我们的框架可以以非常高的概率(约95美元)验证GNN模型的所有权,用于这两项任务。最后,我们实验表明,我们的水标记方法对于两个模型的修改和输入后防御系统防御进行强力。