Supervised federated learning (FL) enables multiple clients to share the trained model without sharing their labeled data. However, potential clients might even be reluctant to label their own data, which could limit the applicability of FL in practice. In this paper, we show the possibility of unsupervised FL whose model is still a classifier for predicting class labels, if the class-prior probabilities are shifted while the class-conditional distributions are shared among the unlabeled data owned by the clients. We propose federation of unsupervised learning (FedUL), where the unlabeled data are transformed into surrogate labeled data for each of the clients, a modified model is trained by supervised FL, and the wanted model is recovered from the modified model. FedUL is a very general solution to unsupervised FL: it is compatible with many supervised FL methods, and the recovery of the wanted model can be theoretically guaranteed as if the data have been labeled. Experiments on benchmark and real-world datasets demonstrate the effectiveness of FedUL. Code is available at https://github.com/lunanbit/FedUL.
翻译:联邦监督学习(FL) 使多个客户能够分享经过培训的模型,而不分享其标签数据。然而,潜在客户甚至可能不愿意贴上自己的数据,这可能会限制FL的实际适用性。在本文中,我们展示了无监督FL的可能性,FL的模型仍然是预测类标签的分类器,如果等级-主要概率变化,而等级-条件分布则由客户拥有的未标签数据共享。我们提议将未经监督的学习(FedUL)联合起来,其中未标签数据被转换为每个客户的代名标签数据,一个经过监督的FL经过修改的模型经过培训,并且从修改的模型中回收通缉的模型。FedUL是一个非常普遍的办法来预测类标签:它与许多受监督的FL方法相兼容,而且对通缉模型的回收在理论上可以保证数据被贴上标签。基准实验和真实世界数据集展示了FDUL的效能。 https://github.com/Flubitrbit https://guth.