Federated learning (FL) enables multiple clients to collaboratively train an accurate global model while protecting clients' data privacy. However, FL is susceptible to Byzantine attacks from malicious participants. Although the problem has gained significant attention, existing defenses have several flaws: the server irrationally chooses malicious clients for aggregation even after they have been detected in previous rounds; the defenses perform ineffectively against sybil attacks or in the heterogeneous data setting. To overcome these issues, we propose MAB-RFL, a new method for robust aggregation in FL. By modelling the client selection as an extended multi-armed bandit (MAB) problem, we propose an adaptive client selection strategy to choose honest clients that are more likely to contribute high-quality updates. We then propose two approaches to identify malicious updates from sybil and non-sybil attacks, based on which rewards for each client selection decision can be accurately evaluated to discourage malicious behaviors. MAB-RFL achieves a satisfying balance between exploration and exploitation on the potential benign clients. Extensive experimental results show that MAB-RFL outperforms existing defenses in three attack scenarios under different percentages of attackers.
翻译:联邦学习(FL)使多个客户能够合作训练准确的全球模型,同时保护客户的数据隐私。然而,FL很容易受到拜占庭恶意参与者的攻击。虽然问题已经引起极大关注,但现有的防御存在若干缺陷:服务器不合理地选择恶意客户进行汇总,即使前几轮已经检测出恶意客户;防御工作没有有效防止周期性袭击或在不同的数据设置中发挥作用。为了克服这些问题,我们建议MAB-RFL(FL)是FL中一种强力整合的新方法。通过模拟客户选择是一个扩大的多臂强盗(MAB)问题,我们提出了一个适应性客户选择战略,以选择更可能提供高质量更新的诚实客户。我们随后提出了两种办法,以识别来自周期性和非周期性袭击的恶意更新,在此基础上可以准确评价对客户选择决定的奖励,以阻止恶意行为。MAB-RFL(RFL)在潜在良性客户的勘探和剥削之间取得令人满意的平衡。广泛的实验结果表明,MAB-RFL(MAB-RFL)在三种攻击情景下,在不同的攻击情景下超越现有防御。