Heating and cooling systems in buildings account for 31% of global energy use, much of which are regulated by Rule Based Controllers (RBCs) that neither maximise energy efficiency nor minimise emissions by interacting optimally with the grid. Control via Reinforcement Learning (RL) has been shown to significantly improve building energy efficiency, but existing solutions require access to building-specific simulators or data that cannot be expected for every building in the world. In response, we show it is possible to obtain emission-reducing policies without such knowledge a priori--a paradigm we call zero-shot building control. We combine ideas from system identification and model-based RL to create PEARL (Probabilistic Emission-Abating Reinforcement Learning) and show that a short period of active exploration is all that is required to build a performant model. In experiments across three varied building energy simulations, we show PEARL outperforms an existing RBC once, and popular RL baselines in all cases, reducing building emissions by as much as 31% whilst maintaining thermal comfort. Our source code is available online via https://enjeeneer.io/projects/pearl/
翻译:建筑物的供暖和冷却系统占全球能源使用量的31%,其中大部分由基于规则的主计长(RBC)管理,后者既不能优化地与电网互动,从而最大限度地提高能源效率,也不能通过最大限度地减少排放。通过强化学习(RL)进行控制,已证明大大提高了建筑能效,但现有的解决方案需要获得建筑专用模拟器或无法预期的世界上每一座建筑都使用的数据。对此,我们表示,在没有事先了解的情况下,有可能获得减少排放的政策,我们称之为“零弹式建筑控制”的范式。我们把系统识别和基于模型的RL的理念结合起来,以创建PEARL(安全授权-减轻强化学习),并显示,要建立一个性能模型,只需要进行短期的积极探索。在三个不同的建筑能源模拟实验中,我们显示PEARL曾经超越了现有的RBC,在所有案例中都超越了流行RL的基线,在维持热安慰的同时将建筑排放量减少多达31 %。我们的源码可通过https://enjeener.io/proal/pal/pal/pall上网查阅。我们的来源代码。