Federated learning (FL) is a privacy-preserving machine learning technique that facilitates collaboration among participants across demographics. FL enables model sharing, while restricting the movement of data. Since FL provides participants with independence over their training data, it becomes susceptible to poisoning attacks. Such collaboration also propagates bias among the participants, even unintentionally, due to different data distribution or historical bias present in the data. This paper proposes an intentional fairness attack, where a client maliciously sends a biased model, by increasing the fairness loss while training, even considering homogeneous data distribution. The fairness loss is calculated by solving an optimization problem for fairness metrics such as demographic parity and equalized odds. The attack is insidious and hard to detect, as it maintains global accuracy even after increasing the bias. We evaluate our attack against the state-of-the-art Byzantine-robust and fairness-aware aggregation schemes over different datasets, in various settings. The empirical results demonstrate the attack efficacy by increasing the bias up to 90\%, even in the presence of a single malicious client in the FL system.
翻译:联邦学习(FL)是一种保护隐私的机器学习技术,它促进了跨人口统计学参与者的协作。FL实现了模型共享,同时限制了数据的移动。由于FL赋予参与者对其训练数据的独立性,因此容易受到投毒攻击。这种协作还会在参与者之间传播偏见,即使是无意的,也可能由于数据分布不同或数据中存在历史偏见而导致。本文提出了一种有意的公平性攻击,其中客户端恶意发送一个有偏模型,通过在训练过程中增加公平性损失来实现,甚至考虑了同质数据分布。公平性损失通过求解公平性指标(如人口统计均等和均衡几率)的优化问题来计算。这种攻击具有隐蔽性且难以检测,因为即使在增加偏见后,它仍能保持全局准确性。我们在不同数据集和各种设置下,评估了我们的攻击对最先进的拜占庭鲁棒和公平感知聚合方案的效果。实证结果表明,即使在FL系统中仅存在单个恶意客户端的情况下,该攻击也能将偏见提高高达90%。