Federated Learning (FL) empowers Industrial Internet of Things (IIoT) with distributed intelligence of industrial automation thanks to its capability of distributed machine learning without any raw data exchange. However, it is rather challenging for lightweight IIoT devices to perform computation-intensive local model training over large-scale deep neural networks (DNNs). Driven by this issue, we develop a communication-computation efficient FL framework for resource-limited IIoT networks that integrates DNN partition technique into the standard FL mechanism, wherein IIoT devices perform local model training over the bottom layers of the objective DNN, and offload the top layers to the edge gateway side. Considering imbalanced data distribution, we derive the device-specific participation rate to involve the devices with better data distribution in more communication rounds. Upon deriving the device-specific participation rate, we propose to minimize the training delay under the constraints of device-specific participation rate, energy consumption and memory usage. To this end, we formulate a joint optimization problem of device scheduling and resource allocation (i.e. DNN partition point, channel assignment, transmit power, and computation frequency), and solve the long-term min-max mixed integer non-linear programming based on the Lyapunov technique. In particular, the proposed dynamic device scheduling and resource allocation (DDSRA) algorithm can achieve a trade-off to balance the training delay minimization and FL performance. We also provide the FL convergence bound for the DDSRA algorithm with both convex and non-convex settings. Experimental results demonstrate the derived device-specific participation rate in terms of feasibility, and show that the DDSRA algorithm outperforms baselines in terms of test accuracy and convergence time.
翻译:联邦学习(FL) 赋予工业互联网(IIoT) 权力,并凭借其分布式机器学习的能力,在不进行任何原始数据交换的情况下进行分布式机器学习,使工业自动化智能得到传播。然而,对于轻型IIoT设备来说,在大型深神经网络(DNN)中进行计算密集的地方模型培训是相当困难的。 受这一问题驱使,我们为资源有限的IIoT网络开发了一个通信-计算高效FL框架,将DNN分区技术纳入标准FL机制,在目标DNN的底层进行本地模型培训,并将顶层从顶层向边缘端端端端端端端端端端端端端端端端端端端端端的机器学习。考虑到数据分布不均匀,我们从设备专用参与率的角度出发,在更大规模深神经神经网络(DRA)中进行计算,在特定装置参与率、能源消耗和记忆使用的限制下,我们为设备调度和资源分配(即DNNU、频道分配、传输、传输频率)的顶端点端点端点端点端点端点、DDDR(也显示以长期最低-DDR标准的升级的升级的升级的进度)的升级,在长期预算安排中,我们可以显示以稳定进行不固定的升级的升级的进度。