Traditionally, AI models are trained on the central cloud with data collected from end devices. This leads to high communication cost, long response time and privacy concerns. Recently Edge empowered AI, namely Edge AI, has been proposed to support AI model learning and deployment at the network edge closer to the data sources. Existing research including federated learning adopts a centralized architecture for model learning where a central server aggregates the model updates from the clients/workers. The centralized architecture has drawbacks such as performance bottleneck, poor scalability and single point of failure. In this paper, we propose a novel decentralized model learning approach, namely E-Tree, which makes use of a well-designed tree structure imposed on the edge devices. The tree structure and the locations and orders of aggregation on the tree are optimally designed to improve the training convergency and model accuracy. In particular, we design an efficient device clustering algorithm, named by KMA, for E-Tree by taking into account the data distribution on the devices as well as the the network distance. Evaluation results show E-Tree significantly outperforms the benchmark approaches such as federated learning and Gossip learning under NonIID data in terms of model accuracy and convergency.
翻译:传统上,AI模型在中央云层上培训,其数据来自终端设备。这导致通信成本高、反应时间长和隐私问题。最近,Edge授权AI,即Edge AI, 提议支持AI 模型学习和在网络边缘更接近数据源的地方部署。现有的研究,包括联合学习,采用了一个中央学习结构用于模型学习的中央结构,中央服务器从客户/工人那里汇总模型更新。中央结构有缺陷,如性能瓶颈、可缩缩缩缩和单一失败点等。在本文中,我们提出了一个新的分散化模式学习方法,即E-Tree,它使边缘设备上强加的精心设计的树结构得以使用。树木结构以及树上集合的位置和顺序的优化设计,是为了改进培训的趋同性和模型准确性。特别是,我们设计了一个高效的设备组合算法,由 KMA 命名为E-Tree,其中考虑到设备上的数据分布以及网络距离。评估结果显示E-Tre 明显地超越了基准方法,例如Federatedlening and Gos Gosils Lon。