To improve the system performance towards the Shannon limit, advanced radio resource management mechanisms play a fundamental role. In particular, scheduling should receive much attention, because it allocates radio resources among different users in terms of their channel conditions and QoS requirements. The difficulties of scheduling algorithms are the tradeoffs need to be made among multiple objectives, such as throughput, fairness and packet drop rate. We propose a smart scheduling scheme based on deep reinforcement learning (DRL). We not only verify the performance gain achieved, but also provide implementation-friend designs, i.e., a scalable neural network design for the agent and a virtual environment training framework. With the scalable neural network design, the DRL agent can easily handle the cases when the number of active users is time-varying without the need to redesign and retrain the DRL agent. Training the DRL agent in a virtual environment offline first and using it as the initial version in the practical usage helps to prevent the system from suffering from performance and robustness degradation due to the time-consuming training. Through both simulations and field tests, we show that the DRL-based smart scheduling outperforms the conventional scheduling method and can be adopted in practical systems.
翻译:为提高系统在香农极限方面的绩效,先进的无线电资源管理机制发挥着根本作用。特别是,日程安排应受到重视,因为它按频道条件和QOS要求在不同用户之间分配无线电资源。时间安排算法的困难在于需要权衡多种目标,如输送量、公平性和投放量。我们提议了一个基于深层强化学习的智能日程安排计划(DRL),我们不仅核查业绩收益,而且还为代理商和虚拟环境培训框架提供可缩放的神经网络设计。在可缩放的神经网络设计下,DRL代理商可以很容易地处理在活跃用户数量是时间变化而无需重新设计和重新配置DRL代理商的情况下出现的情况。我们首先在虚拟环境中对DRL代理商进行离线培训,并将它作为实际使用的初始版本。我们通过模拟和实地测试,显示基于实际的智能日程安排方法能够超越常规形式。