Deep reinforcement learning (DRL) has been increasingly employed to handle the dynamic and complex resource management in network slicing. The deployment of DRL policies in real networks, however, is complicated by heterogeneous cell conditions. In this paper, we propose a novel transfer learning (TL) aided multi-agent deep reinforcement learning (MADRL) approach with inter-agent similarity analysis for inter-cell inter-slice resource partitioning. First, we design a coordinated MADRL method with information sharing to intelligently partition resource to slices and manage inter-cell interference. Second, we propose an integrated TL method to transfer the learned DRL policies among different local agents for accelerating the policy deployment. The method is composed of a new domain and task similarity measurement approach and a new knowledge transfer approach, which resolves the problem of from whom to transfer and how to transfer. We evaluated the proposed solution with extensive simulations in a system-level simulator and show that our approach outperforms the state-of-the-art solutions in terms of performance, convergence speed and sample efficiency. Moreover, by applying TL, we achieve an additional gain over 27% higher than the coordinate MADRL approach without TL.
翻译:深度强化学习(DRL)已被越来越多地用于处理网络切片中的动态和复杂的资源管理。但是,在实际网络中运用DRL政策,由于不同的细胞条件不同而变得复杂。在本文件中,我们建议采用新的传输学习(TL)辅助多剂深度强化学习(MADRL)方法,对跨细胞间切片资源分割进行试样分析(MADRL)方法。首先,我们设计了一个协调的MADRL方法,通过信息分享,将智能分割资源用于切片和管理跨细胞干扰。第二,我们建议采用综合TL方法,将学习过的DRL政策转让给不同的当地代理,以加速政策部署。这个方法由新的领域和任务相似度测量方法和新的知识转移方法组成,解决从谁那里转移和如何转移的问题。我们用系统级模拟器对拟议解决方案进行了广泛的模拟,并表明我们的方法在业绩、趋同速度和样本效率方面超越了最先进的解决方案。此外,通过应用TL,我们取得了比MADRL更高的27%的额外收益。