As a promising distributed machine learning paradigm, Federated Learning (FL) trains a central model with decentralized data without compromising user privacy, which has made it widely used by Artificial Intelligence Internet of Things (AIoT) applications. However, the traditional FL suffers from model inaccuracy since it trains local models using hard labels of data and ignores useful information of incorrect predictions with small probabilities. Although various solutions try to tackle the bottleneck of the traditional FL, most of them introduce significant communication and memory overhead, making the deployment of large-scale AIoT devices a great challenge. To address the above problem, this paper presents a novel Distillation-based Federated Learning (DFL) architecture that enables efficient and accurate FL for AIoT applications. Inspired by Knowledge Distillation (KD) that can increase the model accuracy, our approach adds the soft targets used by KD to the FL model training, which occupies negligible network resources. The soft targets are generated by local sample predictions of each AIoT device after each round of local training and used for the next round of model training. During the local training of DFL, both soft targets and hard labels are used as approximation objectives of model predictions to improve model accuracy by supplementing the knowledge of soft targets. To further improve the performance of our DFL model, we design a dynamic adjustment strategy for tuning the ratio of two loss functions used in KD, which can maximize the use of both soft targets and hard labels. Comprehensive experimental results on well-known benchmarks show that our approach can significantly improve the model accuracy of FL with both Independent and Identically Distributed (IID) and non-IID data.
翻译:作为有希望的分布式机器学习范例,Federal Learning(FL)培训了一个中央模型,其数据分散,但又不损害用户隐私,这使用户隐私。然而,传统FL使用数据硬标签培训当地模型,忽视不正确预测的有用信息,忽视了概率小的错误预测。尽管各种解决方案试图解决传统FL的瓶颈问题,但大多数解决方案都引入了重要的软通信和记忆管理,使得大规模AIOT装置的部署成为一项巨大的挑战。为了解决上述问题,本文展示了一个新的基于Federal Intal Internet(FT)应用应用的基于人工智能智能(AIoT)应用应用软件(AIoT)的网络信息网络信息网络(FL)应用模型不准确性模型(KD)架构,因为知识蒸馏(KD)能够提高模型准确性,我们的方法将KDL使用的软目标添加到FL模式培训中,通过每一轮的当地抽样预测方法产生软目标,并且用于下一轮的模型培训,我们使用软性的AIT工具的软性指标和软性FL的软性指标。在使用FL的模型的模型中,同时,我们用于了软性指标的软性指标的软性指标的精确的模型的模型的精确性目标,用于了使用。