In this paper, we develop a deep learning-based bandwidth allocation policy that is: 1) scalable with the number of users and 2) transferable to different communication scenarios, such as non-stationary wireless channels, different quality-of-service (QoS) requirements, and dynamically available resources. To support scalability, the bandwidth allocation policy is represented by a graph neural network (GNN), with which the number of training parameters does not change with the number of users. To enable the generalization of the GNN, we develop a hybrid-task meta-learning (HML) algorithm that trains the initial parameters of the GNN with different communication scenarios during meta-training. Next, during meta-testing, a few samples are used to fine-tune the GNN with unseen communication scenarios. Simulation results demonstrate that our HML approach can improve the initial performance by 8.79%, and sample efficiency by 73%, compared with existing benchmarks. After fine-tuning, our near-optimal GNN-based policy can achieve close to the same reward with much lower inference complexity compared to the optimal policy obtained using iterative optimization. Numerical results validate that our HML can reduce the computation time by approximately 200 to 2000 times than the optimal iterative algorithm.
翻译:本文提出一种基于深度学习的带宽分配策略,该策略具备以下特性:1)可随用户数量扩展;2)可迁移至不同通信场景,如非平稳无线信道、差异化服务质量(QoS)需求及动态可用资源。为实现可扩展性,带宽分配策略采用图神经网络(GNN)进行建模,其训练参数量不随用户数增加而变化。为提升GNN的泛化能力,我们设计了混合任务元学习(HML)算法,在元训练阶段通过多类通信场景训练GNN的初始参数。随后在元测试阶段,仅需少量样本即可对未见过通信场景下的GNN进行微调。仿真结果表明:相较于现有基准方法,我们的HML方法可将初始性能提升8.79%,样本效率提高73%。经微调后,基于GNN的近最优策略能以远低于迭代优化所得最优策略的推理复杂度,实现与之相近的奖励值。数值结果验证了HML算法可将计算时间缩减至最优迭代算法的约1/200至1/2000。