In contrast to centralized model training that involves data collection, federated learning (FL) enables remote clients to collaboratively train a model without exposing their private data. However, model performance usually degrades in FL due to the heterogeneous data generated by clients of diverse characteristics. One promising strategy to maintain good performance is by limiting the local training from drifting far away from the global model. Previous studies accomplish this by regularizing the distance between the representations learned by the local and global models. However, they only consider representations from the early layers of a model or the layer preceding the output layer. In this study, we introduce FedIntR, which provides a more fine-grained regularization by integrating the representations of intermediate layers into the local training process. Specifically, FedIntR computes a regularization term that encourages the closeness between the intermediate layer representations of the local and global models. Additionally, FedIntR automatically determines the contribution of each layer's representation to the regularization term based on the similarity between local and global representations. We conduct extensive experiments on various datasets to show that FedIntR can achieve equivalent or higher performance compared to the state-of-the-art approaches. Our code is available at https://github.com/YLTun/FedIntR.
翻译:联邦学习(FL)相对于集中式模型训练,不需要数据集中,可以使远程客户端合作训练模型,而不暴露他们的私有数据。然而,FL 中生成的异构数据常常会导致模型性能下降。为了维持良好的性能,一种有前途的策略是限制本地模型训练远离全局模型。先前的研究通过规范化本地和全局模型学习的表示之间的距离来实现这一点。然而,他们只考虑到模型的早期层或输出层之前的层的表示。在本研究中,我们介绍了 FedIntR,它通过将中间层的表示集成到本地训练过程中,提供了更精细的规则化。具体而言,FedIntR 计算一个规范化项,鼓励本地和全局模型的中间层表示之间的接近度。此外,FedIntR 根据本地和全局表示之间的相似性自动确定每个层表示对规则化项的贡献。我们对各种数据集进行了广泛的实验,展示了 FedIntR 相对于最先进的方法可以达到相等或更高的性能。我们的代码可用于https://github.com/YLTun/FedIntR。