Deep neural networks (DNNs) have demonstrated their superiority in practice. Arguably, the rapid development of DNNs is largely benefited from high-quality (open-sourced) datasets, based on which researchers and developers can easily evaluate and improve their learning methods. Since the data collection is usually time-consuming or even expensive, how to protect their copyrights is of great significance and worth further exploration. In this paper, we revisit dataset ownership verification. We find that existing verification methods introduced new security risks in DNNs trained on the protected dataset, due to the targeted nature of poison-only backdoor watermarks. To alleviate this problem, in this work, we explore the untargeted backdoor watermarking scheme, where the abnormal model behaviors are not deterministic. Specifically, we introduce two dispersibilities and prove their correlation, based on which we design the untargeted backdoor watermark under both poisoned-label and clean-label settings. We also discuss how to use the proposed untargeted backdoor watermark for dataset ownership verification. Experiments on benchmark datasets verify the effectiveness of our methods and their resistance to existing backdoor defenses. Our codes are available at \url{https://github.com/THUYimingLi/Untargeted_Backdoor_Watermark}.
翻译:深度神经网络(DNNs)已经在实践中证明了它们的优越性。可以说,DNNs的快速发展在很大程度上得益于高质量(开源)数据集,基于这些数据集,研究人员和开发者可以轻松地评估和改进他们的学习方法。由于数据集的收集通常是耗时甚至昂贵的,如何保护它们的版权具有重大意义,值得进一步探讨。
在本文中,我们重新审视了数据集所有权验证。我们发现,由于以毒性标签为唯一后门水印的定向本质,现有的验证方法在受保护的数据集上训练的DNNs中引入了新的安全风险。为减轻这个问题,本文中我们探讨了未定向后门水印方案,其中异常的模型行为是不确定的。具体地,我们介绍了两种分散度并证明了它们的相关性,基于此设计了在受污染标签和清洁标签设置下的未定向后门水印。我们还讨论了如何使用所提出的未定向后门水印进行数据集所有权验证。对基准数据集的实验验证了我们的方法的有效性以及它们对现有后门防御的抵抗力。我们的代码可在\url{https://github.com/THUYimingLi/Untargeted_Backdoor_Watermark}获得。