Federated learning (FL) has been recognized as one of the most promising solutions for distributed machine learning (ML). In most of the current literature, FL has been studied for supervised ML tasks, in which edge devices collect labeled data. Nevertheless, in many applications, it is impractical to assume existence of labeled data across devices. To this end, we develop a novel methodology, Cooperative Federated unsupervised Contrastive Learning (CF-CL), for FL across edge devices with unlabeled datasets. CF-CL employs local device cooperation where data are exchanged among devices through device-to-device (D2D) communications to avoid local model bias resulting from non-independent and identically distributed (non-i.i.d.) local datasets. CF-CL introduces a push-pull smart data sharing mechanism tailored to unsupervised FL settings, in which, each device pushes a subset of its local datapoints to its neighbors as reserved data points, and pulls a set of datapoints from its neighbors, sampled through a probabilistic importance sampling technique. We demonstrate that CF-CL leads to (i) alignment of unsupervised learned latent spaces across devices, (ii) faster global convergence, allowing for less frequent global model aggregations; and (iii) is effective in extreme non-i.i.d. data settings across the devices.
翻译:联邦学习(FL)被公认为是分布式机器学习(ML)最有希望的解决办法之一。在大多数现有文献中,FL被研究为监督式ML任务,其中边缘装置收集了标签数据。然而,在许多应用中,假设跨设备存在标签数据是不切实际的。为此,我们开发了一种新颖的方法,即FL合作型未经监督的反竞争学习(CF-CL),用于具有未贴标签数据集的边缘装置;CF-CL使用当地设备合作,通过设备对设备(D2D)的通信在设备之间交换数据,以避免由于不独立和同样分布(非i.i.d.)的地方模型偏差。 然而,在许多应用中,CF-CL引入了一种为不受监督的FL设置定制的推式智能数据共享机制(CL-CL),其中每个设备将其本地数据点的一组作为保留式数据点推向邻居,从邻居之间抽取一组数据,通过非稳定性重要取样技术进行数据交换。我们展示的是,CFCFC-CL-C-CL-C-C-CLAxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx