Federated learning (FL) is a privacy-preserving learning paradigm that allows multiple parities to jointly train a powerful machine learning model without sharing their private data. According to the form of collaboration, FL can be further divided into horizontal federated learning (HFL) and vertical federated learning (VFL). In HFL, participants share the same feature space and collaborate on data samples, while in VFL, participants share the same sample IDs and collaborate on features. VFL has a broader scope of applications and is arguably more suitable for joint model training between large enterprises. In this paper, we focus on VFL and investigate potential privacy leakage in real-world VFL frameworks. We design and implement two practical privacy attacks: reverse multiplication attack for the logistic regression VFL protocol; and reverse sum attack for the XGBoost VFL protocol. We empirically show that the two attacks are (1) effective - the adversary can successfully steal the private training data, even when the intermediate outputs are encrypted to protect data privacy; (2) evasive - the attacks do not deviate from the protocol specification nor deteriorate the accuracy of the target model; and (3) easy - the adversary needs little prior knowledge about the data distribution of the target participant. We also show the leaked information is as effective as the raw training data in training an alternative classifier. We further discuss potential countermeasures and their challenges, which we hope can lead to several promising research directions.
翻译:联邦学习(FL)是一种保护隐私的学习模式,它使多种平等能够联合培训强大的机器学习模式,而不必分享其私人数据。根据合作的形式,FL可以进一步分为横向联合学习(HFL)和纵向联合学习(VFL)。在HFL,参与者拥有相同的特征空间,在数据样本上进行合作,而在VFL,参与者共享相同的样本ID,在功能上进行合作。VFL具有更广泛的应用范围,并被认为更适合大型企业之间的联合示范培训。在本文中,我们侧重于VFL,调查真实世界VFL框架中潜在的隐私渗漏。我们设计和实施两种实际隐私攻击:为后勤回归VFL协议反倍增攻击;为XGBoost VFL协议反向袭击。我们从经验上表明,两次攻击是有效的(1) 对手可以成功窃取私人培训数据,即使中间产出被加密以保护数据隐私;(2) 回避----攻击并不偏离协议规格,也不降低目标模型的准确性;(3)我们设计和实施两种实际隐私攻击――我们作为潜在数据传播方向的先期,我们讨论。