Federated learning is a distributed machine learning approach in which a single server and multiple clients collaboratively build machine learning models without sharing datasets on clients. A challenging issue of federated learning is data heterogeneity (i.e., data distributions may differ across clients). To cope with this issue, numerous federated learning methods aim at personalized federated learning and build optimized models for clients. Whereas existing studies empirically evaluated their own methods, the experimental settings (e.g., comparison methods, datasets, and client setting) in these studies differ from each other, and it is unclear which personalized federate learning method achieves the best performance and how much progress can be made by using these methods instead of standard (i.e., non-personalized) federated learning. In this paper, we benchmark the performance of existing personalized federated learning through comprehensive experiments to evaluate the characteristics of each method. Our experimental study shows that (1) there are no champion methods, (2) large data heterogeneity often leads to high accurate predictions, and (3) standard federated learning methods (e.g. FedAvg) with fine-tuning often outperform personalized federated learning methods. We open our benchmark tool FedBench for researchers to conduct experimental studies with various experimental settings.
翻译:联邦学习是一种分布式的机器学习方法,一个单一服务器和多个客户在不分享客户数据集的情况下合作建立机器学习模式,联邦学习的一个棘手问题是数据异质性(即,不同客户的数据分布可能不同)。为了解决这个问题,许多联邦学习方法旨在个人化联合会学习,并为客户建立最佳模式。虽然现有的研究经验评估了自己的方法,但这些研究的实验环境(例如,比较方法、数据集和客户设置)各不相同,而个人化的联邦学习方法往往导致高准确的预测,不清楚哪些个人化的联邦学习方法取得最佳的成绩,以及使用这些方法而不是标准(即,非个人化)联邦化的联邦学习可以取得多大进展。在本文件中,我们通过综合实验来衡量现有个人化联合会学习的绩效,以评价每种方法的特点。我们的实验研究表明:(1)没有支持性方法,(2)大的数据异质性往往导致高准确的预测,(3)标准联邦化学习方法(例如,FedAv,非个人化)通过使用标准的实验性实验方法,经常进行个人化实验性化的实验性实验性研究。