Federated learning (FL) is a promising way to use the computing power of mobile devices while maintaining the privacy of users. Current work in FL, however, makes the unrealistic assumption that the users have ground-truth labels on their devices, while also assuming that the server has neither data nor labels. In this work, we consider the more realistic scenario where the users have only unlabeled data, while the server has some labeled data, and where the amount of labeled data is smaller than the amount of unlabeled data. We call this learning problem semi-supervised federated learning (SSFL). For SSFL, we demonstrate that a critical issue that affects the test accuracy is the large gradient diversity of the models from different users. Based on this, we investigate several design choices. First, we find that the so-called consistency regularization loss (CRL), which is widely used in semi-supervised learning, performs reasonably well but has large gradient diversity. Second, we find that Batch Normalization (BN) increases gradient diversity. Replacing BN with the recently-proposed Group Normalization (GN) can reduce gradient diversity and improve test accuracy. Third, we show that CRL combined with GN still has a large gradient diversity when the number of users is large. Based on these results, we propose a novel grouping-based model averaging method to replace the FedAvg averaging method. Overall, our grouping-based averaging, combined with GN and CRL, achieves better test accuracy than not just a contemporary paper on SSFL in the same settings (>10\%), but also four supervised FL algorithms.
翻译:联邦学习( FL) 是使用移动设备计算能力,同时维护用户隐私的一个很有希望的方法。 然而, FL 目前的工作却不切实际地假设用户在设备上贴有地面真相标签, 同时假设服务器没有数据或标签。 在这项工作中, 我们考虑更现实的假设, 即用户只有未贴标签的数据, 而服务器有一些标签数据, 标签数据的数量比未贴标签数据的数量要小。 我们称这个学习问题半监督的联邦学习( SSFL ) 。 对于 SSLL 来说, 一个影响测试准确性的关键问题是不同用户的模型的大幅梯度多样性。 基于这个假设, 我们调查了几个设计选项。 首先, 我们发现所谓的一致性调整损失( CRL) (CRL ) 被广泛用于半超标的学习, 运行质量差异也相当大。 其次, 我们发现 Batch 正常化( BN) 会增加梯度多样性。 重新将BN 与最近推出的集团的正常化( GN) 联合文件( GN), 我们也可以用高的CR 标准级测试方法来降低 。