Federated learning has emerged as an important distributed learning paradigm, where a server aggregates a global model from many client-trained models while having no access to the client data. Although it is recognized that statistical heterogeneity of the client local data yields slower global model convergence, it is less commonly recognized that it also yields a biased federated global model with a high variance of accuracy across clients. In this work, we aim to provide federated learning schemes with improved fairness. To tackle this challenge, we propose a novel federated learning system that employs zero-shot data augmentation on under-represented data to mitigate statistical heterogeneity and encourage more uniform accuracy performance across clients in federated networks. We study two variants of this scheme, Fed-ZDAC (federated learning with zero-shot data augmentation at the clients) and Fed-ZDAS (federated learning with zero-shot data augmentation at the server). Empirical results on a suite of datasets demonstrate the effectiveness of our methods on simultaneously improving the test accuracy and fairness.
翻译:联邦学习已成为一个重要的分布式学习模式,在这种模式中,服务器从许多客户培训的模型中汇总了一个全球模型,但无法查阅客户数据。虽然人们认识到,当地客户数据的统计异质性导致全球模式趋同速度慢,但不太普遍地认识到,它也产生了偏颇的联邦全球模型,客户的准确性差异很大。在这项工作中,我们的目标是提供联邦学习计划,提高公平性。为了应对这一挑战,我们提议建立一个新型的联邦学习系统,对任职人数不足的数据采用零发数据增强法,以缓解统计异质性,并鼓励在联邦化网络中各客户的准确性表现更加一致。我们研究了这一方案的两种变式,即美联储-ZDAS(用户零发数据增强法)和美联-ZDAS(服务器零发数据增强法)。一套数据集的成果表明我们同时提高测试准确性和公正性的方法的有效性。