Training deep neural networks requires large scale data, which often forces users to work in a distributed or outsourced setting, accompanied with privacy concerns. Split learning framework aims to address this concern by splitting up the model among the client and the server. The idea is that since the server does not have access to client's part of the model, the scheme supposedly provides privacy. We show that this is not true via two novel attacks. (1) We show that an honest-but-curious split learning server, equipped only with the knowledge of the client neural network architecture, can recover the input samples and also obtain a functionally similar model to the client model, without the client being able to detect the attack. (2) Furthermore, we show that if split learning is used naively to protect the training labels, the honest-but-curious server can infer the labels with perfect accuracy. We test our attacks using three benchmark datasets and investigate various properties of the overall system that affect the attacks' effectiveness. Our results show that plaintext split learning paradigm can pose serious security risks and provide no more than a false sense of security.
翻译:培训深层神经网络需要大规模的数据,这些数据往往迫使用户在分布式或外包环境下工作,并伴有隐私问题。分解学习框架旨在通过将模型在客户和服务器之间分离来解决这一关切。想法是,由于服务器无法接触模型的客户部分,因此这个办法据称可以提供隐私。我们通过两次新的袭击表明,这并非事实。 (1) 我们显示,一个只配备客户神经网络结构知识的诚实但充满争议的分解学习服务器,可以回收输入样本,并获得与客户模式功能相似的模式,客户无法检测袭击。 (2) 此外,我们表明,如果将分解学习用于保护培训标签是天真的,诚实但有说服力的服务器可以完全准确地推断标签。我们用三个基准数据集测试我们的攻击,并调查影响袭击有效性的整个系统的各种特性。我们的结果显示,简洁的分解学习模式可以造成严重的安全风险,只能提供虚假的安全感。