Traditional power grid systems have become obsolete under more frequent and extreme natural disasters. Reinforcement learning (RL) has been a promising solution for resilience given its successful history of power grid control. However, most power grid simulators and RL interfaces do not support simulation of power grid under large-scale blackouts or when the network is divided into sub-networks. In this study, we proposed an updated power grid simulator built on Grid2Op, an existing simulator and RL interface, and experimented on limiting the action and observation spaces of Grid2Op. By testing with DDQN and SliceRDQN algorithms, we found that reduced action spaces significantly improve training performance and efficiency. In addition, we investigated a low-rank neural network regularization method for deep Q-learning, one of the most widely used RL algorithms, in this power grid control scenario. As a result, the experiment demonstrated that in the power grid simulation environment, adopting this method will significantly increase the performance of RL agents.
翻译:在更频繁和极端的自然灾害中,传统的电网系统已经过时。强化学习(RL)是抵御能力的一个很有希望的解决方案,因为其成功的电网控制历史是成功的。然而,大多数电网模拟器和RL界面并不支持在大规模停电或网络分为子网络的情况下模拟电网。在这项研究中,我们提议在Grid2Op(现有模拟器和RL界面)上建造最新的电网模拟器,并试验限制Grid2Op的动作和观察空间。通过DDQN和SiceRDQN算法测试,我们发现行动空间缩小将大大改善培训绩效和效率。此外,我们调查了在这种电网控制情景中最广泛使用的电网算法之一,即低级别神经网络的深度学习正规化方法。结果显示,在电网模拟环境中,采用这种方法将大大提高RL代理的性能。