Newton-type methods are popular in federated learning due to their fast convergence. Still, they suffer from two main issues, namely: low communication efficiency and low privacy due to the requirement of sending Hessian information from clients to parameter server (PS). In this work, we introduced a novel framework called FedNew in which there is no need to transmit Hessian information from clients to PS, hence resolving the bottleneck to improve communication efficiency. In addition, FedNew hides the gradient information and results in a privacy-preserving approach compared to the existing state-of-the-art. The core novel idea in FedNew is to introduce a two level framework, and alternate between updating the inverse Hessian-gradient product using only one alternating direction method of multipliers (ADMM) step and then performing the global model update using Newton's method. Though only one ADMM pass is used to approximate the inverse Hessian-gradient product at each iteration, we develop a novel theoretical approach to show the converging behavior of FedNew for convex problems. Additionally, a significant reduction in communication overhead is achieved by utilizing stochastic quantization. Numerical results using real datasets show the superiority of FedNew compared to existing methods in terms of communication costs.
翻译:牛顿型方法在联结学习中很受欢迎,因为它们快速趋同。但是,它们仍然受到两个主要问题的困扰,即:通信效率低,隐私低,因为需要将黑森族信息从客户发送到参数服务器(PS)。在这项工作中,我们引入了名为FedNew的新框架,无需将黑森族信息从客户传递到PS,从而解决瓶颈问题,以提高通信效率。此外,美联储隐藏了梯度信息,并导致与现有最新技术相比,采取隐私保护方法。美联储的核心新理念是引入一个两级框架,在更新逆向海珊级产品时,仅使用一种交替的乘数级方法(ADMM)一步,然后使用牛顿方法进行全球模型更新之间,我们引入了一个新的框架。尽管只有一台ADMM通行证用于在每次循环中接近逆向海珊级产品,但我们开发了一个新的理论方法,以显示美联储在共产问题方面的趋同行为。此外,通过使用真实的通信成本来大幅降低美联运的平价。