To investigate the heterogeneity of federated learning in real-world scenarios, we generalize the classical federated learning to federated hetero-task learning, which emphasizes the inconsistency across the participants in federated learning in terms of both data distribution and learning tasks. We also present B-FHTL, a federated hetero-task learning benchmark consisted of simulation dataset, FL protocols and a unified evaluation mechanism. B-FHTL dataset contains three well-designed federated learning tasks with increasing heterogeneity. Each task simulates the clients with different data distributions and learning tasks. To ensure fair comparison among different FL algorithms, B-FHTL builds in a full suite of FL protocols by providing high-level APIs to avoid privacy leakage, and presets most common evaluation metrics spanning across different learning tasks, such as regression, classification, text generation and etc. Furthermore, we compare the FL algorithms in fields of federated multi-task learning, federated personalization and federated meta learning within B-FHTL, and highlight the influence of heterogeneity and difficulties of federated hetero-task learning. Our benchmark, including the federated dataset, protocols, the evaluation mechanism and the preliminary experiment, is open-sourced at https://github.com/alibaba/FederatedScope/tree/contest/v1.0.
翻译:为了调查在现实世界情景中联邦式学习的异质性,我们推广了传统联邦式学习与联邦式异质性异质性异质性学习,强调在数据分配和学习任务方面,参与者在联邦式学习方面存在不一致之处。我们还提供了B-FHTL,即联邦式异质性接触学习基准,由模拟数据集、FL协议和统一的评估机制组成。B-FHTL数据集包含三个设计完善的开放式学习任务,其异质性日益增加。每项任务都以不同的数据分配和学习任务模拟客户。为了确保不同FL算法之间的公平比较,B-FHTL建立一套完整的FL协议,提供高层次的API以避免隐私渗漏,并预设涵盖不同学习任务(如回归、分类、文本生成等)的最常见的评价标准。此外,我们比较了在联邦式多式任务学习、配制个人化和联邦式个人化化的元数据学习,在B-FFTL内部的HTL内部, 和我们基质化的HTA/实验机制中,包括我们级/联邦式黑质/实验的模型化的数据评估/分析。