We propose an algorithm based on reinforcement learning for solving NP-hard problems on graphs. We combine Graph Isomorphism Networks and the Monte-Carlo Tree Search, which was originally used for game searches, for solving combinatorial optimization on graphs. Similarly to AlphaGo Zero, our method does not require any problem-specific knowledge or labeled datasets (exact solutions), which are difficult to calculate in principle. We show that our method, which is trained by generated random graphs, successfully finds near-optimal solutions for the Maximum Independent Set problem on citation networks. Experiments illustrate that the performance of our method is comparable to SOTA solvers, but we do not require any problem-specific reduction rules, which is highly desirable in practice since collecting hand-crafted reduction rules is costly and not adaptive for a wide range of problems.
翻译:我们提出基于强化学习的算法,以解决图表中的NP-硬问题。 我们将图形形态网络和蒙特-卡洛树搜索(Monte-Carlo树搜索(Monte-Carlo Tree Search)合并在一起,后者最初用于游戏搜索,用于解决图表中的组合优化。 与阿尔法戈零星相似,我们的方法并不要求任何难以在原则上计算的问题特定知识或标签数据集(精确解决方案),我们显示我们的方法(通过生成的随机图表来培训)成功地为引用网络上的最大独立设置问题找到近乎最佳的解决方案。 实验表明,我们的方法的性能与SOTA解答器相似,但我们并不要求任何特定问题的减少规则,而在实践中非常可取,因为收集手工制作的减排规则成本高昂,对广泛的问题没有适应性。