The sixth-generation (6G) mobile networks are expected to feature the ubiquitous deployment of machine learning and AI algorithms at the network edge. With rapid advancements in edge AI, the time has come to realize intelligence downloading onto edge devices (e.g., smartphones and sensors). To materialize this version, we propose a novel technology in this article, called in-situ model downloading, that aims to achieve transparent and real-time replacement of on-device AI models by downloading from an AI library in the network. Its distinctive feature is the adaptation of downloading to time-varying situations (e.g., application, location, and time), devices' heterogeneous storage-and-computing capacities, and channel states. A key component of the presented framework is a set of techniques that dynamically compress a downloaded model at the depth-level, parameter-level, or bit-level to support adaptive model downloading. We further propose a virtualized 6G network architecture customized for deploying in-situ model downloading with the key feature of a three-tier (edge, local, and central) AI library. Furthermore, experiments are conducted to quantify 6G connectivity requirements and research opportunities pertaining to the proposed technology are discussed.
翻译:第六代(6G)移动网络预计在网络边缘广泛部署机器学习和人工智能算法。随着边缘人工智能的快速发展,实现对边缘设备(如智能手机和传感器)的智能下载已经到来。为了实现这个版本,我们在本文中提出一种新的技术,称为原位模型下载,旨在通过从网络中的人工智能库中下载,实现对设备上的人工智能模型的透明和实时替换。其独特之处在于将下载适应于时变情况(如应用程序、位置和时间)、设备异构存储和计算能力以及信道状态。所提出的框架的关键组成部分是一组技术,通过深度级别、参数级别或比特级别的动态压缩下载的模型,以支持自适应模型下载。我们进一步提出了一种虚拟化的6G网络架构,定制了三层(边缘、本地和中央)人工智能库,以部署原位模型下载的关键功能。此外,进行了一些实验来量化6G连接需求,并讨论了与所提出技术相关的研究机会。