During network planning phase, optimal network planning implemented through efficient resource allocation and static traffic demand provisioning in IP-over-elastic optical network (IP-over-EON) is significantly challenging compared with the fixed-grid wavelength division multiplexing (WDM) network due to increased flexibility in IP-over-EON. Mathematical optimization models used for this purpose may not provide solution for large networks due to large computational complexity. In this regard, a greedy heuristic may be used that intuitively selects traffic elements in sequence from static traffic demand matrix and attempts to find the best solution. However, in general, such greedy heuristics offer suboptimal solutions, since appropriate traffic sequence offering the optimal performance is rarely selected. In this regard, we propose a reinforcement learning technique (in particular a Q-learning method), combined with an auxiliary graph (AG)-based energy efficient greedy method to be used for large network planning. The Q-learning method is used to decide the suitable sequence of traffic allocation such that the overall power consumption in the network reduces. In the proposed heuristic, each traffic from the given static traffic demand matrix is successively selected using the Q-learning technique and provisioned using the AG-based greedy method.
翻译:在网络规划阶段,由于IP-超弹性光学网络(IP-over-EON)具有更大的灵活性,因此,通过高效资源分配和静态交通需求提供IP-超弹性光学网络(IP-over-EON)实施的最佳网络规划与固定网格波长分多氧化网络(WDM)网络相比具有相当大的挑战性。由于IP-over-EON的灵活度提高,为此使用的数学优化模型可能无法为大型网络提供解决方案,因为计算复杂程度很大,因此,为此使用的数学优化模式可能无法为大型网络提供解决方案。在这方面,可使用贪婪的惯性惯性惯性惯性模式,即从静态交通需求矩阵中按顺序直接选择交通流量要素,并试图找到最佳的解决方案。然而,一般来说,这种贪婪的超自然现象提供了亚优的解决方案,因为很少选择适当的交通序列提供最佳性能。在这方面,我们提议采用强化学习技术(特别是Q-学习方法),结合基于辅助图形(AG)的节能高效贪婪方法,用于大型网络规划。使用Q-学习方法决定交通分配的适当顺序,以便整个网络电力消耗减少。在拟议的超常态中选择基于固定交通需求的每次交通需求。