Federated learning (FL) allows multiple clients with (private) data to collaboratively train a common machine learning model without sharing their private training data. In-the-wild deployment of FL faces two major hurdles: robustness to poisoning attacks and communication efficiency. To address these concurrently, we propose Federated Supermask Learning (FSL). FSL server trains a global subnetwork within a randomly initialized neural network by aggregating local subnetworks of all collaborating clients. FSL clients share local subnetworks in the form of rankings of network edges; more useful edges have higher ranks. By sharing integer rankings, instead of float weights, FSL restricts the space available to craft effective poisoning updates, and by sharing subnetworks, FSL reduces the communication cost of training. We show theoretically and empirically that FSL is robust by design and also significantly communication efficient; all this without compromising clients' privacy. Our experiments demonstrate the superiority of FSL in real-world FL settings; in particular, (1) FSL achieves similar performances as state-of-the-art FedAvg with significantly lower communication costs: for CIFAR10, FSL achieves same performance as Federated Averaging while reducing communication cost by ~35%. (2) FSL is substantially more robust to poisoning attacks than state-of-the-art robust aggregation algorithms. We have released the code for reproducibility.
翻译:联邦学习联合会(FL)使拥有(私人)数据的多个客户能够合作培训共同机器学习模式,而不必分享其私人培训数据。在部署联邦学习联合会时面临两大障碍:对攻击和通信效率进行毒害的稳健性;同时,我们提议采用FSL(FSL)服务器,通过将所有协作客户的本地子网络集合起来,在随机初始神经网络内培训一个全球子网络。FSL(FSL)客户以网络边缘排名的形式共享本地子网络;更有用的边缘有更高的级别。通过共享整级排名,而不是浮动重量,FSL(FSL)限制了可用于有效中毒更新的空间,通过共享子网络,FSL(FSL)降低了培训的通信成本。我们从理论上和从经验上表明FSL(FSL)通过设计变得稳健健,同时也不损害客户的隐私。我们的实验显示FSL(FSL)在现实世界FL(FL)环境中的优势;特别是,我们(FSL)实现了与最先进的FDA(FS-L)公司(FSL)公司(FS-FSL)公司(FSL)公司(FS-FRS)公司)公司(FS-xL)公司(FAR)公司(FSL)公司)公司(FS-xL)公司)公司)公司(FSRL)公司)公司(在大幅降低快速)(FSDL)(FAR)(FAR)(FSB)(FAR)(FSRFSB)(FAR)(FAR)(FSDAR)(FS)(FAR)(FAR)(FS)(FAR)(FAR)(FAR)(FL)(FL)(FAR)(FA)(FL)(FL)(FL)(FL)(FL)(FL)(FS)(FS)(FAR)(FS)(FAR)(FAR)(FA)(FAR)(FAR)(FAR)(FAR)(FAR))(FAR)(FL)(FL)(FL)(FL)(FL)(FAR)(FS)