Federated Learning (FL) enables machine learning model training on distributed edge devices by aggregating local model updates rather than local data. However, privacy concerns arise as the FL server's access to local model updates can potentially reveal sensitive personal information by performing attacks like gradient inversion recovery. To address these concerns, privacy-preserving methods, such as Homomorphic Encryption (HE)-based approaches, have been proposed. Despite HE's post-quantum security advantages, its applications suffer from impractical overheads. In this paper, we present FedML-HE, the first practical system for efficient HE-based secure federated aggregation that provides a user/device-friendly deployment platform. FL-HE utilizes a novel universal overhead optimization scheme, significantly reducing both computation and communication overheads during deployment while providing customizable privacy guarantees. Our optimized system demonstrates considerable overhead reduction, particularly for large models (e.g., ~10x reduction for HE-federated training of ResNet-50 and ~40x reduction for BERT), demonstrating the potential for scalable HE-based FL deployment.
翻译:随着联邦学习(FL)的兴起,人们可以通过聚合本地模型更新而非本地数据在分布式设备上训练机器学习模型。然而,由于FL服务器对本地模型更新的访问可能会导致隐私泄露,因此出现了隐私保护方法,例如基于同态加密的方法。尽管同态加密具有后量子安全性的优势,但其应用存在着不实际的开销。本文提出了FedML-HE,这是一种高效的基于同态加密的安全联邦聚合系统,并提供用户/设备友好的部署平台。FL-HE采用了一种新颖的通用开销优化方案,大大减少了部署过程中的计算和通信开销,同时提供可定制的隐私保证。我们优化后的系统在大型模型上展示了相当大的开销降低,例如RESNET-50的HE联邦训练可以减少约10倍,BERT的HE联邦训练可以减少约40倍,展示了基于同态加密的FL的可扩展性。