Federated learning (FL) is a distributed learning methodology that allows multiple nodes to cooperatively train a deep learning model, without the need to share their local data. It is a promising solution for telemonitoring systems that demand intensive data collection, for detection, classification, and prediction of future events, from different locations while maintaining a strict privacy constraint. Due to privacy concerns and critical communication bottlenecks, it can become impractical to send the FL updated models to a centralized server. Thus, this paper studies the potential of hierarchical FL in IoT heterogeneous systems and propose an optimized solution for user assignment and resource allocation on multiple edge nodes. In particular, this work focuses on a generic class of machine learning models that are trained using gradient-descent-based schemes while considering the practical constraints of non-uniformly distributed data across different users. We evaluate the proposed system using two real-world datasets, and we show that it outperforms state-of-the-art FL solutions. In particular, our numerical results highlight the effectiveness of our approach and its ability to provide 4-6% increase in the classification accuracy, with respect to hierarchical FL schemes that consider distance-based user assignment. Furthermore, the proposed approach could significantly accelerate FL training and reduce communication overhead by providing 75-85% reduction in the communication rounds between edge nodes and the centralized server, for the same model accuracy.
翻译:联邦学习(FL)是一种分布式学习方法,它使多个节点能够合作培训深层次学习模式,而无需分享其当地数据;它是一个很有希望的解决方案,适用于要求收集密集数据、探测、分类和预测来自不同地点的未来事件的远程监测系统,同时保持严格的隐私限制;由于隐私关切和关键的通信瓶颈,将FL更新模式发送到中央服务器可能变得不切实际。因此,本文件研究IoT混杂系统中等级FL的潜力,并提出在多个边缘节点上用户分配和资源分配的最佳解决方案;特别是,这项工作侧重于利用基于梯度-白种的计划进行培训的通用机器学习模式,同时考虑不同用户之间非统一分布数据的实际限制;我们利用两个真实世界数据集对拟议的系统进行评估,我们表明它比最先进的FL解决方案更不切实际。我们的数字结果突出表明了我们的方法的有效性及其在分类准确度方面提高4-6%的能力,对于等级FL计划而言,在考虑中央级F级用户配置中加速75级的远程用户配置之间通信,我们提出的降低75个端用户配置的方法。