The label distribution skew induced data heterogeniety has been shown to be a significant obstacle that limits the model performance in federated learning, which is particularly developed for collaborative model training over decentralized data sources while preserving user privacy. This challenge could be more serious when the participating clients are in unstable circumstances and dropout frequently. Previous work and our empirical observations demonstrate that the classifier head for classification task is more sensitive to label skew and the unstable performance of FedAvg mainly lies in the imbalanced training samples across different classes. The biased classifier head will also impact the learning of feature representations. Therefore, maintaining a balanced classifier head is of significant importance for building a better global model. To this end, we propose a simple yet effective framework by introducing a prior-calibrated softmax function for computing the cross-entropy loss and a prototype-based feature augmentation scheme to re-balance the local training, which are lightweight for edge devices and can facilitate the global model aggregation. The improved model performance over existing baselines in the presence of non-IID data and client dropout is demonstrated by conducting extensive experiments on benchmark classification tasks.
翻译:标签分布扭曲导致的数据差异性被证明是一个重大障碍,限制了联邦学习模式的示范性业绩,而联邦学习模式是特别为针对分散的数据源进行合作模式培训而开发的,同时保护用户隐私。当参与的客户处于不稳定环境和经常辍学时,这一挑战可能更为严峻。以前的工作和我们的经验观察表明,分类任务的分类负责人对标签扭曲和FedAvg的不稳定性表现更加敏感,主要在于不同类别的培训样本不平衡。有偏见的分类头也会影响特征表现的学习。因此,保持平衡的分类头对于建立一个更好的全球模型非常重要。为此,我们提出一个简单而有效的框架,在计算跨热带损失时引入一个事先校准的软体积功能,以及一个基于原型特征增强计划,以重新平衡当地培训,因为边缘装置的重量较轻,而且能够促进全球模型汇总。在非IID数据和客户辍学情况下对现有基线的改进模型性表现通过对基准分类任务进行广泛的试验来证明。</s>