Distributed (or Federated) learning enables users to train machine learning models on their very own devices, while they share only the gradients of their models usually in a differentially private way (utility loss). Although such a strategy provides better privacy guarantees than the traditional centralized approach, it requires users to blindly trust a centralized infrastructure that may also become a bottleneck with the increasing number of users. In this paper, we design and implement P4L: a privacy preserving peer-to-peer learning system for users to participate in an asynchronous, collaborative learning scheme without requiring any sort of infrastructure or relying on differential privacy. Our design uses strong cryptographic primitives to preserve both the confidentiality and utility of the shared gradients, a set of peer-to-peer mechanisms for fault tolerance and user churn, proximity and cross device communications. Extensive simulations under different network settings and ML scenarios for three real-life datasets show that P4L provides competitive performance to baselines, while it is resilient to different poisoning attacks. We implement P4L and experimental results show that the performance overhead and power consumption is minimal (less than 3mAh of discharge).
翻译:分布式(或联邦)学习使用户能够用自己的设备培训机器学习模型,而他们通常只以不同私人方式分享其模型的梯度(功能损失),虽然这种战略比传统的中央化方法提供更好的隐私保障,但要求用户盲目信任中央基础设施,这种基础设施也可能成为用户数量不断增加的瓶颈。在本文中,我们设计和实施P4L:一个保护隐私的同侪学习系统,使用户能够参与一个不同步的协作学习计划,而不需要任何基础设施或依赖差异性隐私。我们的设计使用强大的加密原始技术来维护共享梯度的保密性和效用,这是一套对用户的容忍度和用户热量、近距离和交叉设备通信的同侪-同侪机制。在不同网络环境中的广泛模拟和三个真实生命数据集的ML情景表明,P4L为用户提供了竞争性的基线性工作,同时能够抵御不同的中毒攻击。我们实施的P4L和实验结果表明,性压载力和电量消耗是最低的(少于3毫米排放量 )。</s>