This paper investigates the feasibility of learning good representation space with unlabeled client data in the federated scenario. Existing works trivially inherit the supervised federated learning methods, which does not apply to the model heterogeneity and has the potential risk of privacy exposure. To tackle the problems above, we first identify that self-supervised contrastive local training is more robust against the non-i.i.d.-ness than the traditional supervised learning paradigm. Then we propose a novel federated self-supervised contrastive learning framework FLESD that supports architecture-agnostic local training and communication-efficient global aggregation. At each round of communication, the server first gathers a fraction of the clients' inferred similarity matrices on a public dataset. Then FLESD ensembles the similarity matrices and trains the global model via similarity distillation. We verify the effectiveness of our proposed framework by a series of empirical experiments and show that FLESD has three main advantages over the existing methods: it handles the model heterogeneity, is less prone to privacy leak, and is more communication-efficient. We will release the code of this paper in the future.
翻译:本文探讨在联盟式假设情景中以未贴标签的客户数据学习良好代表空间的可行性。 现有的工作微乎其微地继承了受监督的联邦学习方法, 这种方法不适用于模型异质性, 并具有潜在的隐私暴露风险。 为了解决上述问题, 我们首先发现, 自我监督的对比性本地培训比传统受监督的学习模式更加有力。 我们通过一系列经验实验来验证我们提议的框架的有效性, 并表明 FLESD比现有方法有三大优势: 它处理模型异质性, 比较不易发生隐私泄漏, 并且通信效率更高。 我们将在将来发布该文件的代码 。