The Traveling Salesman Problem (TSP) is the most popular and most studied combinatorial problem, starting with von Neumann in 1951. It has driven the discovery of several optimization techniques such as cutting planes, branch-and-bound, local search, Lagrangian relaxation, and simulated annealing. The last five years have seen the emergence of promising techniques where (graph) neural networks have been capable to learn new combinatorial algorithms. The main question is whether deep learning can learn better heuristics from data, i.e. replacing human-engineered heuristics? This is appealing because developing algorithms to tackle efficiently NP-hard problems may require years of research, and many industry problems are combinatorial by nature. In this work, we propose to adapt the recent successful Transformer architecture originally developed for natural language processing to the combinatorial TSP. Training is done by reinforcement learning, hence without TSP training solutions, and decoding uses beam search. We report improved performances over recent learned heuristics with an optimal gap of 0.004% for TSP50 and 0.39% for TSP100.
翻译:旅游推销员问题(TSP)是1951年从冯纽曼开始,最受欢迎和研究最多的组合问题。它促使人们发现了一些优化技术,如切割机、分支和约束、局部搜索、拉格兰加放松和模拟肛门。过去五年中出现了一些有希望的技术,使(绘图)神经网络能够学习新的组合算法。主要问题是深层次学习能否从数据中更好地学到超自然学,即取代人造的超自然学?这是很有吸引力的,因为开发算法以有效解决NP-硬性问题可能需要多年的研究,而许多工业问题是自然的组合问题。在这项工作中,我们提议将原先为自然语言处理而开发的最近成功的变形器结构调整到组合式TSP。培训是通过强化学习完成的,因此没有TSP培训解决方案,解码用途是搜索。我们报告最近学习的超自然学表现有所改善,TSP50和TSP100最佳差距为0.004%和0.39%。