Federated learning (FL) is becoming a popular paradigm for collaborative learning over distributed, private datasets owned by non-trusting entities. FL has seen successful deployment in production environments, and it has been adopted in services such as virtual keyboards, auto-completion, item recommendation, and several IoT applications. However, FL comes with the challenge of performing training over largely heterogeneous datasets, devices, and networks that are out of the control of the centralized FL server. Motivated by this inherent setting, we make a first step towards characterizing the impact of device and behavioral heterogeneity on the trained model. We conduct an extensive empirical study spanning close to 1.5K unique configurations on five popular FL benchmarks. Our analysis shows that these sources of heterogeneity have a major impact on both model performance and fairness, thus sheds light on the importance of considering heterogeneity in FL system design.
翻译:联邦学习(FL)正在成为在分布式、非信任实体拥有的私人数据集方面开展合作学习的流行范例。FL在生产环境中成功部署,并在虚拟键盘、自动完成、项目建议和若干IoT应用等服务中采用。然而,FL面临挑战,要对大部分不同数据集、装置和超出中央FL服务器控制的网络进行培训。受这种内在环境的驱动,我们迈出了第一步,将装置和行为差异性对经过培训的模式的影响定性。我们开展了一项广泛的经验研究,在五种受欢迎的FL基准上覆盖近1.5K的独特配置。我们的分析表明,这些差异性来源对模型性能和公平性都有重大影响,从而说明了在FL系统设计中考虑差异性的重要性。