Federated learning (FL) enables distributed participants to collectively learn a strong global model without sacrificing their individual data privacy. Mainstream FL approaches require each participant to share a common network architecture and further assume that data are are sampled IID across participants. However, in real-world deployments participants may require heterogeneous network architectures; and the data distribution is almost certainly non-uniform across participants. To address these issues we introduce FedH2L, which is agnostic to both the model architecture and robust to different data distributions across participants. In contrast to approaches sharing parameters or gradients, FedH2L relies on mutual distillation, exchanging only posteriors on a shared seed set between participants in a decentralized manner. This makes it extremely bandwidth efficient, model agnostic, and crucially produces models capable of performing well on the whole data distribution when learning from heterogeneous silos.
翻译:联邦学习(FL)使分布式参与者能够在不牺牲个人数据隐私的情况下集体学习强大的全球模型,而不必牺牲个人数据隐私。主流FL方法要求每个参与者共享一个共同的网络架构,并进一步假设数据是跨参与者的抽样国际开发数据。然而,在现实世界部署中,参与者可能需要不同的网络架构;而且数据分布几乎肯定不是跨参与者的统一。为了解决这些问题,我们引入了FedH2L(FedH2L)(FedH2L)(FedH2L)(FedH2L)(FedH2L)(Fed ), 与共享参数或梯度不同,FedH2L(F) 依靠的是相互蒸馏,以分散方式在参与者之间共享的种子集上只交换子集。这使得它非常高效的带宽、模型的不可知性,并且关键地生成模型,在学习多式硅时能够很好地完成整个数据分布。