Neural approaches for combinatorial optimization (CO) equip a learning mechanism to discover powerful heuristics for solving complex real-world problems. While neural approaches capable of high-quality solutions in a single shot are emerging, state-of-the-art approaches are often unable to take full advantage of the solving time available to them. In contrast, hand-crafted heuristics perform highly effective search well and exploit the computation time given to them, but contain heuristics that are difficult to adapt to a dataset being solved. With the goal of providing a powerful search procedure to neural CO approaches, we propose simulation-guided beam search (SGBS), which examines candidate solutions within a fixed-width tree search that both a neural net-learned policy and a simulation (rollout) identify as promising. We further hybridize SGBS with efficient active search (EAS), where SGBS enhances the quality of solutions backpropagated in EAS, and EAS improves the quality of the policy used in SGBS. We evaluate our methods on well-known CO benchmarks and show that SGBS significantly improves the quality of the solutions found under reasonable runtime assumptions.
翻译:组合式优化的神经方法(CO) 装备了一种学习机制,以发现强大的超自然理论解决复杂的现实世界问题。虽然一种能够以一针见血地找到高质量解决方案的神经方法正在出现,但最先进的神经方法往往无法充分利用可供它们使用的解答时间。相比之下,手工制作的超自然技术进行非常有效的搜索,利用赋予它们的计算时间,但含有难以适应正在解决的数据集的超自然学。为了提供强大的神经CO方法搜索程序,我们提议模拟制导波音搜索(SGBS),在固定宽树搜索中审查候选人解决方案,发现神经网络学习政策和模拟(滚动)都具有希望。我们进一步将SGBS与高效的积极搜索(EAS)相结合,使SGBS能够提高在电子AS中反向调整的解决方案的质量,而EAS则提高在SGBS中所使用的政策的质量。我们评估了我们众所周知的CO基准方法,并表明在合理时间下发现的解决办法的质量。