This paper proposes a new algorithm, referred to as GMAB, that combines concepts from the reinforcement learning domain of multi-armed bandits and random search strategies from the domain of genetic algorithms to solve discrete stochastic optimization problems via simulation. In particular, the focus is on noisy large-scale problems, which often involve a multitude of dimensions as well as multiple local optima. Our aim is to combine the property of multi-armed bandits to cope with volatile simulation observations with the ability of genetic algorithms to handle high-dimensional solution spaces accompanied by an enormous number of feasible solutions. For this purpose, a multi-armed bandit framework serves as a foundation, where each observed simulation is incorporated into the memory of GMAB. Based on this memory, genetic operators guide the search, as they provide powerful tools for exploration as well as exploitation. The empirical results demonstrate that GMAB achieves superior performance compared to benchmark algorithms from the literature in a large variety of test problems. In all experiments, GMAB required considerably fewer simulations to achieve similar or (far) better solutions than those generated by existing methods. At the same time, GMAB's overhead with regard to the required runtime is extremely small due to the suggested tree-based implementation of its memory. Furthermore, we prove its convergence to the set of global optima as the simulation effort goes to infinity.
翻译:本文提出一种新的算法,称为“GMAB”,将多臂强盗强化学习领域的概念与基因算法领域随机搜索战略的概念结合起来,通过模拟解决离散的随机优化问题。特别是,重点是吵闹的大规模问题,这些问题往往涉及多个层面和多重局部选择。我们的目标是将多臂强盗的特性与基因算法处理高维解决方案空间的能力结合起来,同时辅之以大量可行的解决办法。为此,多臂强盗框架是一个基础,每个观察到的模拟都被纳入GMAB的记忆中。基于这一记忆,基因操作者指导搜索,因为它们为勘探和开发提供了强有力的工具。经验结果显示,与大量测试问题的文献基准算法相比,GMAB取得优异性的工作表现。在所有实验中,GMAB所需要的模拟比现有方法产生的要少得多或(远)更好的解决办法。与此同时,GAB的顶部与所需的全球模范努力相比,我们所建议的全球模范努力的趋近度是小的。