Federated learning (FL) enables distributed participants to collaboratively learn a global model without revealing their private data to each other. Recently, vertical FL, where the participants hold the same set of samples but with different features, has received increased attention. This paper first presents one label inference attack method to investigate the potential privacy leakages of the vertical logistic regression model. Specifically, we discover that the attacker can utilize the residue variables, which are calculated by solving the system of linear equations constructed by local dataset and the received decrypted gradients, to infer the privately owned labels. To deal with this, we then propose three protection mechanisms, e.g., additive noise mechanism, multiplicative noise mechanism, and hybrid mechanism which leverages local differential privacy and homomorphic encryption techniques, to prevent the attack and improve the robustness of the vertical logistic regression. model. Experimental results show that both the additive noise mechanism and the multiplicative noise mechanism can achieve efficient label protection with only a slight drop in model testing accuracy, furthermore, the hybrid mechanism can achieve label protection without any testing accuracy degradation, which demonstrates the effectiveness and efficiency of our protection techniques
翻译:联邦学习(FL)使分布式参与者能够合作学习一个全球模型,而没有向对方透露其私人数据。最近,垂直FL,即参与者持有相同的样本,但具有不同特点的垂直FL得到越来越多的关注。本文首先介绍了一种标签推断攻击方法,以调查纵向物流回归模型潜在的隐私泄漏。具体地说,我们发现攻击者可以利用残余变量,通过解决由当地数据集和收到的解密梯度所构建的线性方程式系统来计算,推断私有标签。为此,我们然后提议三种保护机制,例如添加噪音机制、倍增噪音机制和混合机制,以及利用本地差异隐私和同式加密技术的混合机制,以防止攻击并提高垂直物流回归的稳健性。模型实验结果显示,添加噪音机制和多复制噪音机制都能够实现高效的标签保护,而模型测试精度则略有下降。此外,混合机制可以在不测试准确性降解的情况下实现标签保护,这表明我们保护技术的有效性和效率。