Federated learning allows clients to collaboratively learn statistical models while keeping their data local. Federated learning was originally used to train a unique global model to be served to all clients, but this approach might be sub-optimal when clients' local data distributions are heterogeneous. In order to tackle this limitation, recent personalized federated learning methods train a separate model for each client while still leveraging the knowledge available at other clients. In this work, we exploit the ability of deep neural networks to extract high quality vectorial representations (embeddings) from non-tabular data, e.g., images and text, to propose a personalization mechanism based on local memorization. Personalization is obtained by interpolating a collectively trained global model with a local $k$-nearest neighbors (kNN) model based on the shared representation provided by the global model. We provide generalization bounds for the proposed approach in the case of binary classification, and we show on a suite of federated datasets that this approach achieves significantly higher accuracy and fairness than state-of-the-art methods.
翻译:联邦学习使客户能够合作学习统计模型,同时将其数据保存在本地。联邦学习最初用于培训一个独特的全球模型,供所有客户使用,但当客户的当地数据分布不一时,这一方法可能并不理想。为了应对这一限制,最近个人化的联邦学习方法为每个客户培训了一个单独的模型,同时仍在利用其他客户的知识。在这项工作中,我们利用深神经网络的能力,从非表层数据(例如图像和文本)中提取高质量的矢量表(组合),以提出一个基于当地记忆化的个人化机制。通过根据全球模型提供的共同代表制,对集体培训的全球模型进行相互推算,从而获得一个以当地美元为最远的邻居(kNN)为单位的组合式全球模型。我们为二元分类的拟议方法提供了概括性界限,我们用一套节制数据集展示了这一方法比国家先进方法更准确、更公平得多的方法。