Federated learning (FL) is a distributed machine learning approach involving multiple clients collaboratively training a shared model. Such a system has the advantage of more training data from multiple clients, but data can be non-identically and independently distributed (non-i.i.d.). Privacy and integrity preserving features such as differential privacy (DP) and robust aggregation (RA) are commonly used in FL. In this work, we show that on common deep learning tasks, the performance of FL models differs amongst clients and situations, and FL models can sometimes perform worse than local models due to non-i.i.d. data. Secondly, we show that incorporating DP and RA degrades performance further. Then, we conduct an ablation study on the performance impact of different combinations of common personalization approaches for FL, such as finetuning, mixture-of-experts ensemble, multi-task learning, and knowledge distillation. It is observed that certain combinations of personalization approaches are more impactful in certain scenarios while others always improve performance, and combination approaches are better than individual ones. Most clients obtained better performance with combined personalized FL and recover from performance degradation caused by non-i.i.d. data, DP, and RA.
翻译:联邦学习(FL)是一种分布式的机器学习方法,涉及多个客户合作培训一个共享模式。这样的系统具有来自多个客户的更多培训数据的优势,但数据可以非识别和独立地进行(非i.i.d.)传播。隐私和完整性保护特征,如不同的隐私(DP)和强力聚合(RA)等隐私和完整性保护特征在FL中常用。在这项工作中,我们表明,在共同的深层次学习任务中,FL模型在客户和各种情况中的表现不同,FL模型有时由于非i.d.数据而比当地模型更差。第二,我们表明,纳入DP和RA可以进一步降低业绩。然后,我们就FL通用个性化方法的不同组合,如微调、专家混合组合组合、多功能学习和知识蒸馏等的性能效果,进行了反差研究。我们发现,个人化方法的某些组合在某些情景中影响更大,而其他模式则总是改进性能和组合方法优于单个模型。大多数客户的绩效与个人化的DPL和RA数据相结合。