Federated learning introduces a novel approach to training machine learning (ML) models on distributed data while preserving user's data privacy. This is done by distributing the model to clients to perform training on their local data and computing the final model at a central server. To prevent any data leakage from the local model updates, various works with focus on secure aggregation for privacy preserving federated learning have been proposed. Despite their merits, most of the existing protocols still incur high communication and computation overhead on the participating entities and might not be optimized to efficiently handle the large update vectors for ML models. In this paper, we present E-seaML, a novel secure aggregation protocol with high communication and computation efficiency. E-seaML only requires one round of communication in the aggregation phase and it is up to 318x and 1224x faster for the user and the server (respectively) as compared to its most efficient counterpart. E-seaML also allows for efficiently verifying the integrity of the final model by allowing the aggregation server to generate a proof of honest aggregation for the participating users. This high efficiency and versatility is achieved by extending (and weakening) the assumption of the existing works on the set of honest parties (i.e., users) to a set of assisting nodes. Therefore, we assume a set of assisting nodes which assist the aggregation server in the aggregation process. We also discuss, given the minimal computation and communication overhead on the assisting nodes, how one could assume a set of rotating users to as assisting nodes in each iteration. We provide the open-sourced implementation of E-seaML for public verifiability and testing.
翻译:联邦学习(Federated learning)是一种新颖的方法,它能够在分布式数据上训练机器学习(ML)模型,并保护用户的数据隐私。它通过将模型分发给客户端来进行训练,并在中央服务器上计算最终模型,从而实现这一目标。为了防止来自本地模型更新的任何数据泄漏,各种旨在保护隐私的联邦学习的安全聚合协议相继提出。虽然这些协议有很多优点,但大多数现有的协议仍会对参与实体产生较高的通信和计算负担,并且可能无法有效地处理大型模型更新向量。本文提出了E-seaML,这是一种高效的安全聚合协议。在聚合阶段,E-seaML仅需要一轮通信,与效率最高的对手相比,对用户和服务器的效率分别提高了318倍和1224倍。E-seaML还允许通过允许聚合服务器生成有关参与用户的证明,以高效验证最终模型的完整性。通过扩展(和削弱)现有作品在诚实方(即用户)的集合上的假设,从而实现这种高效和多功能的目的。因此,我们假设一组协助节点来辅助聚合服务器进行聚合。我们还讨论了在每个迭代中如何假设一组轮换用户作为辅助节点,以及在用户的最小计算和通信负载上如何假设这一点。我们为公共可验证性和测试提供了E-seaML的开放源代码实现。