In classical federated learning, the clients contribute to the overall training by communicating local updates for the underlying model on their private data to a coordinating server. However, updating and communicating the entire model becomes prohibitively expensive when resource-constrained clients collectively aim to train a large machine learning model. Split learning provides a natural solution in such a setting, where only a small part of the model is stored and trained on clients while the remaining large part of the model only stays at the servers. However, the model partitioning employed in split learning introduces a significant amount of communication cost. This paper addresses this issue by compressing the additional communication using a novel clustering scheme accompanied by a gradient correction method. Extensive empirical evaluations on image and text benchmarks show that the proposed method can achieve up to $490\times$ communication cost reduction with minimal drop in accuracy, and enables a desirable performance vs. communication trade-off.
翻译:在传统的联邦学习中,客户通过向协调服务器通报其私人数据基础模型的当地更新情况,为总体培训作出贡献;然而,如果受资源限制的客户共同致力于培训大型机器学习模式,则整个模型的更新和传播费用极高,因为受资源限制的客户共同致力于培训大型机器学习模式;在这种环境下,分散学习是一种自然的解决办法,只有一小部分模型储存和对客户进行培训,而其余大部分模型只留在服务器;但是,分离学习中使用的模型分割带来了大量的通信费用;然而,在分离学习中使用的模型分割带来了大量的通信费用;本文通过使用一种新的集群计划,加上梯度校正方法,压缩额外通信,从而解决这一问题。关于图像和文本基准的广泛经验评估表明,拟议方法可以实现最多490美元的时间的通信费用减少,而准确性下降最小,并能够实现理想的绩效与通信交易。