Large-scale neural networks possess considerable expressive power. They are well-suited for complex learning tasks in industrial applications. However, large-scale models pose significant challenges for training under the current Federated Learning (FL) paradigm. Existing approaches for efficient FL training often leverage model parameter dropout. However, manipulating individual model parameters is not only inefficient in meaningfully reducing the communication overhead when training large-scale FL models, but may also be detrimental to the scaling efforts and model performance as shown by recent research. To address these issues, we propose the Federated Opportunistic Block Dropout (FedOBD) approach. The key novelty is that it decomposes large-scale models into semantic blocks so that FL participants can opportunistically upload quantized blocks, which are deemed to be significant towards training the model, to the FL server for aggregation. Extensive experiments evaluating FedOBD against five state-of-the-art approaches based on multiple real-world datasets show that it reduces the overall communication overhead by more than 70% compared to the best performing baseline approach, while achieving the highest test accuracy. To the best of our knowledge, FedOBD is the first approach to perform dropout on FL models at the block level rather than at the individual parameter level.
翻译:大型神经网络拥有相当的显微力,它们非常适合工业应用中复杂的学习任务。然而,大型模型对在目前联邦学习(FL)范式下的培训构成重大挑战。现有的高效FL培训方法往往会影响模型参数的退出。然而,在培训大型FL模型时,操纵个人模型参数不仅在有意义地减少通信间接费用方面效率低下,而且可能不利于最近的研究显示的扩大努力和模型性能。为了解决这些问题,我们提议采用联邦机会区块丢弃(Fedobunistic Blockout) 方法。关键的新颖之处是,它将大型模型分解成语义区块,这样FL参与者可以随机地上传四分区块,这些区块被认为对于培训模型很重要,对于FL服务器汇总十分重要。根据多个真实世界数据集对FOBD的五种最先进方法进行广泛的评估表明,它比第一个最佳的基线方法减少了70 %以上的总体通信间接费用,同时达到最高测试精确度。在FLD标准水平上,在FOBD一级,最佳的参数上是最低水平。