Federated Learning (FL) is a decentralized machine learning protocol that allows a set of participating agents to collaboratively train a model without sharing their data. This makes FL particularly suitable for settings where data privacy is desired. However, it has been observed that the performance of FL is closely tied with the local data distributions of agents. Particularly, in settings where local data distributions vastly differ among agents, FL performs rather poorly with respect to the centralized training. To address this problem, we hypothesize the reasons behind the performance degradation, and develop some techniques to address these reasons accordingly. In this work, we identify four simple techniques that can improve the performance of trained models without incurring any additional communication overhead to FL, but rather, some light computation overhead either on the client, or the server-side. In our experimental analysis, combination of our techniques improved the validation accuracy of a model trained via FL by more than 12% with respect to our baseline. This is about 5% less than the accuracy of the model trained on centralized data.
翻译:联邦学习组织(FL)是一个分散式的机器学习协议,它使一组参与机构能够在不分享数据的情况下合作培训模型,从而使FL特别适合需要数据隐私的环境。然而,据观察,FL的性能与当地代理机构的数据分布密切相关。特别是在当地代理机构的数据分布差异很大的情况下,FL在集中化培训方面表现很差。为了解决这一问题,我们假设性能退化背后的原因,并开发一些技术来相应解决这些问题。在这项工作中,我们确定了四种简单技术,可以改进经过培训的模型的性能,而不会给FL带来额外的通信间接费用,而是在客户或服务器方面进行一些轻量计算。在我们实验分析中,我们的技术结合使通过FL培训的模型在基线方面提高了12%以上。这比在集中化数据方面受过培训的模型的准确率低约5%。