In the traditional distributed machine learning scenario, the user's private data is transmitted between clients and a central server, which results in significant potential privacy risks. In order to balance the issues of data privacy and joint training of models, federated learning (FL) is proposed as a particular distributed machine learning procedure with privacy protection mechanisms, which can achieve multi-party collaborative computing without revealing the original data. However, in practice, FL faces a variety of challenging communication problems. This review seeks to elucidate the relationship between these communication issues by methodically assessing the development of FL communication research from three perspectives: communication efficiency, communication environment, and communication resource allocation. Firstly, we sort out the current challenges existing in the communications of FL. Second, we have collated FL communications-related papers and described the overall development trend of the field based on their logical relationship. Ultimately, we discuss the future directions of research for communications in FL.
翻译:在传统的分布式机器学习设想中,用户的私人数据在客户和中央服务器之间传输,这造成了潜在的重大隐私风险;为了平衡数据隐私问题和模型联合培训,提议将联合学习(FL)作为一种特定的分布式机器学习程序,与隐私保护机制相结合,这种程序可以在不披露原始数据的情况下实现多党合作计算,但在实践上,FL面临各种具有挑战性的通信问题;这一审查力求从三个角度,即通信效率、通信环境和通信资源分配,对FL通信研究的发展进行系统评估,从而阐明这些通信问题之间的关系;首先,我们整理了FL通信中目前存在的挑战;第二,我们整理了与FL通信有关的文件,并根据这些文件的逻辑关系,介绍了该领域的总体发展趋势;最后,我们讨论了FL通信的未来研究方向。