End-to-end training of neural network solvers for combinatorial optimization problems such as the Travelling Salesman Problem is intractable and inefficient beyond a few hundreds of nodes. While state-of-the-art Machine Learning approaches perform closely to classical solvers when trained on trivially small sizes, they are unable to generalize the learnt policy to larger instances of practical scales. Towards leveraging transfer learning to solve large-scale TSPs, this paper identifies inductive biases, model architectures and learning algorithms that promote generalization to instances larger than those seen in training. Our controlled experiments provide the first principled investigation into such zero-shot generalization, revealing that extrapolating beyond training data requires rethinking the neural combinatorial optimization pipeline, from network layers and learning paradigms to evaluation protocols.
翻译:对神经网络解决问题的终端到终端培训,如旅行推销员问题等组合优化问题,在数百个节点之外是难以解决的、效率低下的。尽管最先进的机器学习方法在接受小小小规模的培训时与古典解决方案密切配合,但是它们无法将所学政策推广到更大的实际规模。为了利用转移学习解决大规模TSP问题,本文件确定了诱导偏见、模型架构和学习算法,这些方法能够促进普及到比培训中看到的更大的例子。 我们的控制实验为这种零光化提供了第一项原则性调查,揭示了超越培训数据的外推法需要重新思考神经组合优化管道,从网络层和学习模式到评价协议。