Federated learning (FL) is a recently developed area of machine learning, in which the private data of a large number of distributed clients is used to develop a global model under the coordination of a central server without explicitly exposing the data. The standard FL strategy has a number of significant bottlenecks including large communication requirements and high impact on the clients' resources. Several strategies have been described in the literature trying to address these issues. In this paper, a novel scheme based on the notion of "model growing" is proposed. Initially, the server deploys a small model of low complexity, which is trained to capture the data complexity during the initial set of rounds. When the performance of such a model saturates, the server switches to a larger model with the help of function-preserving transformations. The model complexity increases as more data is processed by the clients, and the overall process continues until the desired performance is achieved. Therefore, the most complex model is broadcast only at the final stage in our approach resulting in substantial reduction in communication cost and client computational requirements. The proposed approach is tested extensively on three standard benchmarks and is shown to achieve substantial reduction in communication and client computation while achieving comparable accuracy when compared to the current most effective strategies.
翻译:联邦学习(FL)是一个最近开发的机器学习领域,在这个领域,大量分布的客户的私人数据被用来在中央服务器的协调下开发一个全球模型,而没有明确地暴露数据; 标准的FL战略有一些重大瓶颈,包括大量的通信要求和对客户资源的巨大影响; 文献中介绍了试图解决这些问题的若干战略; 本文提出了一个基于“模式增长”概念的新计划; 最初,服务器部署一个小型的低复杂性模型,经过培训,以在最初的回合中捕捉数据的复杂性; 当模型饱和度运行时,服务器转换为更大的模型,帮助功能保留转换; 模型复杂性随着客户处理更多的数据而增加,整个过程持续到预期的绩效实现; 因此,最复杂的模型仅在我们方法的最后阶段播出,导致通信成本和客户计算要求的大幅下降; 拟议的方法在三个标准基准上进行了广泛测试,显示在与目前最有效的战略相比,通信和客户计算达到可比的准确度,同时实现了通信和客户计算。