In collaborative learning settings like federated learning, curious parities might be honest but are attempting to infer other parties' private data through inference attacks while malicious parties might manipulate the learning process for their own purposes through backdoor attacks. However, most existing works only consider the federated learning scenario where data are partitioned by samples (HFL). The feature-partitioned federated learning (VFL) can be another important scenario in many real-world applications. Attacks and defenses in such scenarios are especially challenging when the attackers and the defenders are not able to access the features or model parameters of other participants. Previous works have only shown that private labels can be reconstructed from per-sample gradients. In this paper, we first show that private labels can be reconstructed when only batch-averaged gradients are revealed, which is against the common presumption. In addition, we show that a passive party in VFL can even replace its corresponding labels in the active party with a target label through a gradient-replacement attack. To defend against the first attack, we introduce a novel technique termed confusional autoencoder (CoAE), based on autoencoder and entropy regularization. We demonstrate that label inference attacks can be successfully blocked by this technique while hurting less main task accuracy compared to existing methods. Our CoAE technique is also effective in defending the gradient-replacement backdoor attack, making it an universal and practical defense strategy with no change to the original VFL protocol. We demonstrate the effectiveness of our approaches under both two-party and multi-party VFL settings. To the best of our knowledge, this is the first systematic study to deal with label inference and backdoor attacks in the feature-partitioned federated learning framework.
翻译:在合作学习环境中,比如联谊学习,好奇的平价可能是诚实的,但试图通过推断攻击来推断其他各方的私人数据,而恶意的各方则可能通过幕后攻击为其自身目的操纵学习过程。然而,大多数现有作品只考虑由样本(HFL)对数据进行分割的联谊学习情景。特异部分的联谊学习(VFL)可能是许多现实应用中的另一个重要情景。当攻击者和捍卫者无法接触其他参与者的特征或模型参数时,这种情景中的攻击和防御尤其具有挑战性。以前的工作仅表明,私人标签可以从每升梯子中为其自身目的操纵学习过程。在本文中,我们首先展示的是,当仅仅分批平均梯子显示数据(HFL)数据时,可以重建私人标签。此外,VFLFL的被动方甚至可以在活跃的一方中用一个目标标签取代其相应的标签,通过梯度-更深度攻击。为了防御第一次攻击,我们引入了一个新的技术,将混乱的自定义的自定义的自译器(CAEE), 以更低的自制的自制的自制的自制的自制的自制式攻击策略,而更精确的自制的自制的自制的自制的自制的自制的自制的自制的自制的自制的自制的自制式的自制式的自制式的自制式的自制式的自制的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式的自制式式式的自制式的自制式式式式式式式的自制式式式式式式式式式式的自制式式式的自制式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式式的自制式的自制式式式式式式式式式式式式式式式式式式式式式式