Federated learning (FL) has emerged as a promising privacy-preserving distributed machine learning framework recently. It aims at collaboratively learning a shared global model by performing distributed training locally on edge devices and aggregating local models into a global one without centralized raw data sharing in the cloud server. However, due to the large local data heterogeneities (Non-I.I.D. data) across edge devices, the FL may easily obtain a global model that can produce more shifted gradients on local datasets, thereby degrading the model performance or even suffering from the non-convergence during training. In this paper, we propose a novel FL training framework, dubbed Fed-FSNet, using a properly designed Fuzzy Synthesizing Network (FSNet) to mitigate the Non-I.I.D. FL at-the-source. Concretely, we maintain an edge-agnostic hidden model in the cloud server to estimate a less-accurate while direction-aware inversion of the global model. The hidden model can then fuzzily synthesize several mimic I.I.D. data samples (sample features) conditioned on only the global model, which can be shared by edge devices to facilitate the FL training towards faster and better convergence. Moreover, since the synthesizing process involves neither access to the parameters/updates of local models nor analyzing individual local model outputs, our framework can still ensure the privacy of FL. Experimental results on several FL benchmarks demonstrate that our method can significantly mitigate the Non-I.I.D. issue and obtain better performance against other representative methods.
翻译:联邦学习(FL)最近已成为一个充满希望的隐私保护分布式机器学习框架,它的目的是通过在当地对边缘设备进行分布式培训,将地方模型整合成一个全球模型,在云端服务器上没有集中的原始数据共享;然而,由于在边缘设备之间有大量的当地数据差异(非I.I.D.数据),FL很容易获得一个全球模型,可以在本地数据集上产生更偏移的梯度,从而降低模型的性能,甚至因培训过程中的非趋同性差而遭受痛苦。在本文件中,我们提议了一个新型FL培训框架,称为FFD-FSNet,使用一个设计得当的Fuzy Synsize化网络(FSNet)来减少非I.D.FL.(非I.D.数据共享数据共享数据共享数据共享性能),具体地说,我们在云层服务器上维持一个边缘性隐蔽模型,以估计一个不太精确的模型,同时识别全球模型的反向性能。 隐藏式模型随后可以对若干MI.I.D.D.