In this paper, to deal with the heterogeneity in federated learning (FL) systems, a knowledge distillation (KD) driven training framework for FL is proposed, where each user can select its neural network model on demand and distill knowledge from a big teacher model using its own private dataset. To overcome the challenge of train the big teacher model in resource limited user devices, the digital twin (DT) is exploit in the way that the teacher model can be trained at DT located in the server with enough computing resources. Then, during model distillation, each user can update the parameters of its model at either the physical entity or the digital agent. The joint problem of model selection and training offloading and resource allocation for users is formulated as a mixed integer programming (MIP) problem. To solve the problem, Q-learning and optimization are jointly used, where Q-learning selects models for users and determines whether to train locally or on the server, and optimization is used to allocate resources for users based on the output of Q-learning. Simulation results show the proposed DT-assisted KD framework and joint optimization method can significantly improve the average accuracy of users while reducing the total delay.
翻译:在本文中,为了应对联合学习(FL)系统中的异质性,提议了一个由知识蒸馏(KD)驱动的FL培训框架,供每个用户根据需求选择其神经网络模型,并用自己的私人数据集从一个大型教师模型中提取知识。为了克服在资源有限的用户设备中培训大型教师模型的挑战,数字双胞胎(DT)正在利用该教师模型,使位于服务器的DT培训有足够的计算资源。然后,在模型蒸馏过程中,每个用户都可以在物理实体或数字代理机构更新其模型参数。模型选择和培训脱载和用户资源分配的共同问题被作为一种混合整数编程(MIP)问题提出。为了解决问题,在Q学习为用户选择模型并确定是在当地还是利用服务器培训,同时优化用于根据Q学习产出为用户分配资源。模拟结果显示拟议的DT辅助KD框架和联合优化方法的总体精确度,同时大幅降低平均用户的延迟度。</s>