This paper proposes a method for wireless network optimization applicable to tuning cell parameters that impact the performance of the adjusted cell and the surrounding neighboring cells. The method relies on multiple reinforcement learning agents that share a common policy and include information from neighboring cells in the state and reward. In order not to impair network performance during the first steps of learning, agents are pre-trained during an earlier phase of offline learning, in which an initial policy is obtained using feedback from a static network simulator and considering a wide variety of scenarios. Finally, agents can wisely tune the cell parameters of a test network by suggesting small incremental changes to slowly steer the network toward an optimal configuration. Agents propose optimal changes using the experience gained with the simulator in the pre-training phase, but also continue to learn from current network readings after each change. The results show how the proposed approach significantly improves the performance gains already provided by expert system-based methods when applied to remote antenna tilt optimization. Additional gains are also seen when comparing the proposed approach with a similar method in which the state and reward do not include information from neighboring cells.
翻译:本文建议了一种适用于调整单元格参数的无线网络优化方法,该方法将影响经调整的单元格和周围相邻单元格的性能。该方法依赖于多个强化学习剂,它们共享共同政策,包括来自州内邻近细胞的信息和奖赏。为了在学习的最初阶段不破坏网络的性能,代理商在离线学习的早期阶段先接受过培训,其中利用静态网络模拟器的反馈获得初始政策,并考虑多种设想。最后,代理商可以明智地调整测试网络的细胞参数,建议小的渐进变化,以缓慢引导网络向最佳配置方向发展。代理商建议采用在培训前阶段与模拟器取得的经验进行最佳变革,但还在每次变化后继续从当前的网络读取经验。结果显示,拟议的方法如何在应用远程天线倾斜优化时大大改进以专家系统为基础的方法已经提供的性能收益。在将拟议方法与类似方法进行比较时还看到额外收益,即状态和奖励办法不包括来自邻近细胞的信息。</s>