Federated Learning (FL) enables multiple parties to distributively train a ML model without revealing their private datasets. However, it assumes trust in the centralized aggregator which stores and aggregates model updates. This makes it prone to gradient tampering and privacy leakage by a malicious aggregator. Malicious parties can also introduce backdoors into the joint model by poisoning the training data or model gradients. To address these issues, we present BEAS, the first blockchain-based framework for N-party FL that provides strict privacy guarantees of training data using gradient pruning (showing improved differential privacy compared to existing noise and clipping based techniques). Anomaly detection protocols are used to minimize the risk of data-poisoning attacks, along with gradient pruning that is further used to limit the efficacy of model-poisoning attacks. We also define a novel protocol to prevent premature convergence in heterogeneous learning environments. We perform extensive experiments on multiple datasets with promising results: BEAS successfully prevents privacy leakage from dataset reconstruction attacks, and minimizes the efficacy of poisoning attacks. Moreover, it achieves an accuracy similar to centralized frameworks, and its communication and computation overheads scale linearly with the number of participants.
翻译:联邦学习联合会(FL)使多个缔约方能够在不透露其私人数据集的情况下对ML模型进行分配性培训。然而,它相信存储和汇总模型更新的中央集成器,从而容易被恶意聚合器造成梯度篡改和隐私泄漏。恶意缔约方还可以通过毒害培训数据或模型梯度,将后门引入联合模型。为解决这些问题,我们介绍了BEAS,这是N-党FL的第一个基于块链的框架,它为使用梯度剪裁剪(显示与现有噪音和剪裁剪法相比的隐私有改进)的培训数据提供了严格的隐私保障。异常检测协议被用来最大限度地减少数据渗透攻击的风险,同时使用梯度剪切线来进一步限制模型渗透攻击的功效。我们还制定了新的协议,以防止多种学习环境中出现过早的趋同。我们在多个数据集上进行了广泛的实验,并取得了大有希望的结果:BEAS成功地防止了数据集重建攻击的隐私泄漏,并最大限度地减少了中毒攻击的功效。此外,它实现了类似于中央框架的准确性参与者,以及其通信和间接计算的数字。