In the traditional distributed machine learning scenario, the user's private data is transmitted between nodes and a central server, which results in great potential privacy risks. In order to balance the issues of data privacy and joint training of models, federated learning (FL) is proposed as a special distributed machine learning with a privacy protection mechanism, which can realize multi-party collaborative computing without revealing the original data. However, in practice, FL faces many challenging communication problems. This review aims to clarify the relationship between these communication problems, and focus on systematically analyzing the research progress of FL communication work from three perspectives: communication efficiency, communication environment, and communication resource allocation. Firstly, we sort out the current challenges existing in the communications of FL. Secondly, we have compiled articles related to FL communications, and then describe the development trend of the entire field guided by the logical relationship between them. Finally, we point out the future research directions for communications in FL.
翻译:在传统的分散式机器学习设想中,用户的私人数据在节点和中央服务器之间传输,这造成了巨大的潜在隐私风险;为了平衡数据隐私和模型联合培训的问题,建议联合学习(FL)作为一种特殊的分散式机器学习和隐私保护机制,可以实现多党合作计算,而不必披露原始数据;然而,在实践中,FL面临许多具有挑战性的通信问题;这一审查旨在澄清这些通信问题之间的关系,并侧重于从三个角度系统分析FL通信工作的研究进展:通信效率、通信环境和通信资源分配。首先,我们整理了FL通信中目前存在的挑战。第二,我们汇编了与FL通信有关的文章,然后描述了以它们之间的逻辑关系为指导的整个领域的发展趋势。最后,我们指出了FL通信的未来研究方向。