Federated learning (FL) enables multiple parties to collaboratively train a machine learning model without sharing their data; rather, they train their own model locally and send updates to a central server for aggregation. Depending on how the data is distributed among the participants, FL can be classified into Horizontal (HFL) and Vertical (VFL). In VFL, the participants share the same set of training instances but only host a different and non-overlapping subset of the whole feature space. Whereas in HFL, each participant shares the same set of features while the training set is split into locally owned training data subsets. VFL is increasingly used in applications like financial fraud detection; nonetheless, very little work has analyzed its security. In this paper, we focus on robustness in VFL, in particular, on backdoor attacks, whereby an adversary attempts to manipulate the aggregate model during the training process to trigger misclassifications. Performing backdoor attacks in VFL is more challenging than in HFL, as the adversary i) does not have access to the labels during training and ii) cannot change the labels as she only has access to the feature embeddings. We present a first-of-its-kind clean-label backdoor attack in VFL, which consists of two phases: a label inference and a backdoor phase. We demonstrate the effectiveness of the attack on three different datasets, investigate the factors involved in its success, and discuss countermeasures to mitigate its impact.
翻译:联邦学习(FL)使多个参与方在不共享数据的情况下共同训练机器学习模型;他们在本地训练自己的模型,然后将更新发送到中央服务器进行聚合。根据数据在参与方之间的分布方式,FL可以分为水平联邦学习(HFL)和垂直联邦学习(VFL)。在VFL中,参与方共享相同的训练实例,但只托管整个特征空间的不同且不重叠的子集。而在HFL中,每个参与方共享相同的特征,而训练集被拆分为本地拥有的训练数据子集。VFL越来越多地应用于诸如金融欺诈检测之类的应用中; 然而,很少有研究分析其安全性。在本文中,我们关注VFL中的鲁棒性,特别是后门攻击,即攻击方试图在训练过程中操纵聚合模型以触发误分类。在VFL中进行后门攻击比在HFL中更具挑战性,因为攻击者 i)在训练期间没有访问标签,ii)无法更改标签,因为她只能访问特征嵌入。我们提出了一种独特的干净标签后门攻击,该攻击包括两个阶段:标签推断和后门阶段。我们证明了该攻击对三个不同数据集的有效性,研究了成功的因素,并讨论了减轻其影响的对策。