To facilitate the emerging applications in 5G networks, mobile network operators will provide many network functions in terms of control and prediction. Recently, they have recognized the power of machine learning (ML) and started to explore its potential to facilitate those network functions. Nevertheless, the current ML models for network functions are often derived in an offline manner, which is inefficient due to the excessive overhead for transmitting a huge volume of dataset to remote ML training clouds and failing to provide the incremental learning capability for the continuous model updating. As an alternative solution, we propose Cocktail, an incremental learning framework within a reference 5G network architecture. To achieve cost efficiency while increasing trained model accuracy, an efficient online data scheduling policy is essential. To this end, we formulate an online data scheduling problem to optimize the framework cost while alleviating the data skew issue caused by the capacity heterogeneity of training workers from the long-term perspective. We exploit the stochastic gradient descent to devise an online asymptotically optimal algorithm, including two optimal policies based on novel graph constructions for skew-aware data collection and data training. Small-scale testbed and large-scale simulations validate the superior performance of our proposed framework.
翻译:为了便利5G网络的新兴应用,移动网络运营商将在控制和预测方面提供许多网络功能;最近,他们认识到机器学习(ML)的力量,并开始探索其促进这些网络功能的潜力;然而,目前网络功能的ML模型往往以离线方式产生,由于向远程ML培训云传播大量数据集的间接费用过多,未能为连续更新模型提供增量学习能力,因此效率低下;作为一种替代解决办法,我们提议Cocktail,即一个参考5G网络架构内的递增学习框架;为了在提高经过培训的模型准确性的同时实现成本效益,有效的在线数据排期政策至关重要;为此,我们制定了在线数据排期问题,以优化框架成本,同时从长期角度减轻培训工作者能力异质引起的数据扭曲问题;我们利用随机梯级梯级梯落,设计一种在线零点最佳算法,包括基于用于Skew-awa数据收集和数据培训的新型图形构造的两个最佳政策;小规模测试台和大规模模拟,验证我们拟议的高级模拟框架的绩效。