Research on deep reinforcement learning (DRL) based production scheduling (PS) has gained a lot of attention in recent years, primarily due to the high demand for optimizing scheduling problems in diverse industry settings. Numerous studies are carried out and published as stand-alone experiments that often vary only slightly with respect to problem setups and solution approaches. The programmatic core of these experiments is typically very similar. Despite this fact, no standardized and resilient framework for experimentation on PS problems with DRL algorithms could be established so far. In this paper, we introduce schlably, a Python-based framework that provides researchers a comprehensive toolset to facilitate the development of PS solution strategies based on DRL. schlably eliminates the redundant overhead work that the creation of a sturdy and flexible backbone requires and increases the comparability and reusability of conducted research work.
翻译:近年来,对基于深度强化学习(DRL)的生产时间安排(PS)的研究引起了许多关注,这主要是因为对优化不同行业的时间安排问题的需求很大,许多研究作为独立实验进行和出版,在问题设置和解决方案方法方面往往略有不同,这些实验的方案核心通常非常相似。尽管如此,迄今尚无法建立标准化和有弹性的框架来试验DRL算法(PS)的问题。在本文件中,我们引入了一个基于Python的框架,为研究人员提供了一种全面的工具,促进制定以DRL为基础的PS解决方案战略。可以肯定地消除了建立坚固和灵活的骨干所需的多余的间接费用,提高了研究工作的可比性和可重复性。</s>