Multi-objective evolutionary algorithms (MOEAs) are widely used to solve multi-objective optimization problems. The algorithms rely on setting appropriate parameters to find good solutions. However, this parameter tuning could be very computationally expensive in solving non-trial (combinatorial) optimization problems. This paper proposes a framework that integrates MOEAs with adaptive parameter control using Deep Reinforcement Learning (DRL). The DRL policy is trained to adaptively set the values that dictate the intensity and probability of mutation for solutions during optimization. We test the proposed approach with a simple benchmark problem and a real-world, complex warehouse design and control problem. The experimental results demonstrate the advantages of our method in terms of solution quality and computation time to reach good solutions. In addition, we show the learned policy is transferable, i.e., the policy trained on a simple benchmark problem can be directly applied to solve the complex warehouse optimization problem, effectively, without the need for retraining.
翻译:多目标进化算法(MOEAs)被广泛用于解决多目标优化问题。算法依靠设定适当的参数来找到良好的解决方案。然而,这一参数调控在解决非工(combinator)优化问题时可能非常昂贵。本文件提出了一个框架,将MOEAs与适应性参数控制相结合,使用深强化学习(DRL) 。DRL政策经过培训,可以适应性地设定决定优化期间解决方案变异强度和概率的值。我们用简单的基准问题和现实世界、复杂的仓库设计和控制问题来测试拟议方法。实验结果显示了我们方法在解决方案质量和计算时间以达成良好解决方案方面的优势。此外,我们展示了所学的政策是可转让的,即,经过简单基准问题培训的政策可以直接用于解决复杂的仓库优化问题,而无需再培训。