A well-trained DNN model can be regarded as an intellectual property (IP) of the model owner. To date, many DNN IP protection methods have been proposed, but most of them are watermarking based verification methods where model owners can only verify their ownership passively after the copyright of DNN models has been infringed. In this paper, we propose an effective framework to actively protect the DNN IP from infringement. Specifically, we encrypt the DNN model's parameters by perturbing them with well-crafted adversarial perturbations. With the encrypted parameters, the accuracy of the DNN model drops significantly, which can prevent malicious infringers from using the model. After the encryption, the positions of encrypted parameters and the values of the added adversarial perturbations form a secret key. Authorized user can use the secret key to decrypt the model. Compared with the watermarking methods which only passively verify the ownership after the infringement occurs, the proposed method can prevent infringement in advance. Moreover, compared with most of the existing active DNN IP protection methods, the proposed method does not require additional training process of the model, which introduces low computational overhead. Experimental results show that, after the encryption, the test accuracy of the model drops by 80.65%, 81.16%, and 87.91% on Fashion-MNIST, CIFAR-10, and GTSRB, respectively. Moreover, the proposed method only needs to encrypt an extremely low number of parameters, and the proportion of the encrypted parameters of all the model's parameters is as low as 0.000205%. The experimental results also indicate that, the proposed method is robust against model fine-tuning attack and model pruning attack. Moreover, for the adaptive attack where attackers know the detailed steps of the proposed method, the proposed method is also demonstrated to be robust.
翻译:训练有素的 DNN 模型可以被视为模型所有者的知识产权( IP) 。 到目前, 许多 DNN 模型保护方法已经提出了许多 DNN 参数保护方法, 但大多数都是基于水的核查方法, 模型所有者只能在DNN 模型的版权被侵犯后被动地验证其所有权。 在本文中, 我们提议了一个有效的框架, 积极保护 DNN 模型不受侵犯。 具体地说, 我们用精心设计的对抗性攻击的触摸来加密 DNN 模型的参数。 到现在, DNN 模型的精度显著下降, 从而防止恶意侵犯者使用模型。 在加密后, 加密参数的位置和添加的对抗性反扰动值形成一个秘密密钥。 授权用户可以使用秘密的密钥来解密该模型。 与仅仅在侵犯行为发生后被动地核查所有权的模型提议方法相比, 模式可以防止提前侵权。 此外, 与大多数现有的 DNN IP IP 保护方法相比, 拟议的方法的精确性会大大下降,, 防止恶意侵犯者使用该模型的精确性 。 也不需要再进行更多的培训过程, 。 在模型的模型中, 测试后, GRER 的精确的精确度测试中, 。