Federated learning is a distributed learning technique where machine learning models are trained on client devices in which the local training data resides. The training is coordinated via a central server which is, typically, controlled by the intended owner of the resulting model. By avoiding the need to transport the training data to the central server, federated learning improves privacy and efficiency. But it raises the risk of model theft by clients because the resulting model is available on every client device. Even if the application software used for local training may attempt to prevent direct access to the model, a malicious client may bypass any such restrictions by reverse engineering the application software. Watermarking is a well-known deterrence method against model theft by providing the means for model owners to demonstrate ownership of their models. Several recent deep neural network (DNN) watermarking techniques use backdooring: training the models with additional mislabeled data. Backdooring requires full access to the training data and control of the training process. This is feasible when a single party trains the model in a centralized manner, but not in a federated learning setting where the training process and training data are distributed among several client devices. In this paper, we present WAFFLE, the first approach to watermark DNN models trained using federated learning. It introduces a retraining step at the server after each aggregation of local models into the global model. We show that WAFFLE efficiently embeds a resilient watermark into models incurring only negligible degradation in test accuracy (-0.17%), and does not require access to training data. We also introduce a novel technique to generate the backdoor used as a watermark. It outperforms prior techniques, imposing no communication, and low computational (+3.2%) overhead.
翻译:联邦学习是一种分布式学习技术,通过当地培训数据所在的客户设备对机器学习模式进行培训。培训通过中央服务器进行协调,通常由该模型的预定所有人控制。通过避免将培训数据传送到中央服务器,联合会学习提高了隐私和效率。但它增加了客户盗窃模式的风险,因为每个客户设备都有由此产生的模式。即使用于当地培训的应用软件可能试图防止直接接触模型,恶意客户也可能通过反向设计应用软件而绕过任何此类限制。水标记是一种众所周知的威慑方法,通过提供模型所有人展示其模型所有权的手段,防止模式盗窃。最近的一些深层神经网络(DNN)水标记技术使用后门:用额外的错误标签数据培训模型。后门要求完全使用培训数据并对培训过程进行控制。单方仅以集中方式对模型进行培训,而不是将培训过程和培训数据在几个客户设备之间进行配置的联邦化学习环境。在本文中,我们将一个经过培训的模型引入了内部数据库。我们使用了一个内部数据模型,然后将一个SNFLEFF升级的方法引入了当地数据库。