The performance of federated learning systems is bottlenecked by communication costs and training variance. The communication overhead problem is usually addressed by three communication-reduction techniques, namely, model compression, partial device participation, and periodic aggregation, at the cost of increased training variance. Different from traditional distributed learning systems, federated learning suffers from data heterogeneity (since the devices sample their data from possibly different distributions), which induces additional variance among devices during training. Various variance-reduced training algorithms have been introduced to combat the effects of data heterogeneity, while they usually cost additional communication resources to deliver necessary control information. Additionally, data privacy remains a critical issue in FL, and thus there have been attempts at bringing Differential Privacy to this framework as a mediator between utility and privacy requirements. This paper investigates the trade-offs between communication costs and training variance under a resource-constrained federated system theoretically and experimentally, and how communication reduction techniques interplay in a differentially private setting. The results provide important insights into designing practical privacy-aware federated learning systems.
翻译:联合学习系统的性能受到通信成本和培训差异的制约。通信间接问题通常通过三种通信减少技术来解决,即模型压缩、部分设备参与和定期汇总,其代价是培训差异的增加。不同于传统的分布式学习系统,联合学习有数据差异(因为设备抽样其数据可能不同分布),这在培训期间造成各种设备之间出现更多的差异。采用了各种差异式培训算法,以消除数据差异性的影响,而通常它们需要额外的通信资源来提供必要的控制信息。此外,数据隐私仍然是FL的关键问题,因此,有人试图将差异隐私作为通用性和隐私要求之间的调和者纳入这一框架。本文调查了在资源限制的封闭式系统下,通信成本与培训差异之间的权衡问题,以及在理论和实验上如何在差异式的私人环境下减少通信技术的相互作用。结果为设计实用的隐私意识强化学习系统提供了重要见解。