Federated Learning (FL) has become a popular distributed learning paradigm that involves multiple clients training a global model collaboratively in a data privacy-preserving manner. However, the data samples usually follow a long-tailed distribution in the real world, and FL on the decentralized and long-tailed data yields a poorly-behaved global model severely biased to the head classes with the majority of the training samples. To alleviate this issue, decoupled training has recently been introduced to FL, considering it has achieved promising results in centralized long-tailed learning by re-balancing the biased classifier after the instance-balanced training. However, the current study restricts the capacity of decoupled training in federated long-tailed learning with a sub-optimal classifier re-trained on a set of pseudo features, due to the unavailability of a global balanced dataset in FL. In this work, in order to re-balance the classifier more effectively, we integrate the local real data with the global gradient prototypes to form the local balanced datasets, and thus re-balance the classifier during the local training. Furthermore, we introduce an extra classifier in the training phase to help model the global data distribution, which addresses the problem of contradictory optimization goals caused by performing classifier re-balancing locally. Extensive experiments show that our method consistently outperforms the existing state-of-the-art methods in various settings.
翻译:联邦学习联合会(FL)已经成为一个广受欢迎的分布式学习模式,它涉及多个客户以数据保密的方式合作培训全球模式;然而,数据样本通常在现实世界中经过长期细致的分发,而关于分散的和长期详细的数据的FL, 产生一种对头类有严重偏差的全球模式,与大多数培训样本严重偏向。为了缓解这一问题,最近向FL引入了脱钩培训,因为通过在对有偏向的分类员进行重新平衡的培训,在集中化的长期扩展学习中取得了有希望的成果。然而,目前的研究限制了以非优化的分类员为分流的混合长期学习进行分解培训的能力,因为没有在FL中建立全球平衡的数据集。 为了更有效地重新平衡分类员,我们最近向FL引入了脱钩培训,将当地真实数据与全球梯度原型数据整合为地方平衡数据集,从而帮助在地方培训中重新平衡分类员。 此外,我们引入了一套不精细的分类方法,通过不断升级的升级的系统,在本地培训中推行了一种不固定的升级的系统,从而将数据升级地展示了一种升级的系统。