In federated learning, each participant trains its local model with its own data and a global model is formed at a trusted server by aggregating model updates coming from these participants. Since the server has no effect and visibility on the training procedure of the participants to ensure privacy, the global model becomes vulnerable to attacks such as data poisoning and model poisoning. Although many defense algorithms have recently been proposed to address these attacks, they often make strong assumptions that do not agree with the nature of federated learning, such as Non-IID datasets. Moreover, they mostly lack comprehensive experimental analyses. In this work, we propose a defense algorithm called BARFED that does not make any assumptions about data distribution, update similarity of participants, or the ratio of the malicious participants. BARFED mainly considers the outlier status of participant updates for each layer of the model architecture based on the distance to the global model. Hence, the participants that do not have any outlier layer are involved in model aggregation. We perform extensive experiments on many grounds and show that the proposed approach provides a robust defense against different attacks.
翻译:在联合学习中,每个参与者用自己的数据来培训自己的本地模型,并且在一个可靠的服务器上通过汇总来自这些参与者的模型更新而形成一个全球模型。由于服务器对参与者的隐私培训程序没有任何影响和可见度,因此全球模型很容易受到数据中毒和模型中毒等攻击。虽然最近提出了许多国防算法来应对这些攻击,但它们往往作出与联合会学习的性质不相符的强烈假设,如非二维数据集。此外,它们大多缺乏全面的实验分析。在这项工作中,我们建议采用称为BARFED的国防算法,即BARFED,不就数据分配、参与者的类似性或恶意参与者的比例作出任何假设。BARFED主要考虑参与者在与全球模型距离的基础上对每一层模型结构进行更新的外部状况。因此,在模型集集中涉及没有外部层的参与者。我们在许多方面进行了广泛的实验,并表明拟议的方法提供了抵御不同攻击的有力防御。