End-to-end training of neural network solvers for graph combinatorial optimization problems such as the Travelling Salesperson Problem (TSP) have seen a surge of interest recently, but remain intractable and inefficient beyond graphs with few hundreds of nodes. While state-of-the-art learning-driven approaches for TSP perform closely to classical solvers when trained on trivially small sizes, they are unable to generalize the learnt policy to larger instances at practical scales. This work presents an end-to-end neural combinatorial optimization pipeline that unifies several recent papers in order to identify the inductive biases, model architectures and learning algorithms that promote generalization to instances larger than those seen in training. Our controlled experiments provide the first principled investigation into such zero-shot generalization, revealing that extrapolating beyond training data requires rethinking the neural combinatorial optimization pipeline, from network layers and learning paradigms to evaluation protocols. Additionally, we analyze recent advances in deep learning for routing problems through the lens of our pipeline and provide new directions to stimulate future research.
翻译:对用于图形组合优化问题的神经网络终端到终端培训,例如旅行销售商问题(TSP)最近引起了人们的极大兴趣,但依然难以解决,效率低下,超过了有几百个节点的图表。 TSP最先进的学习驱动方法在接受小小小尺寸的培训时,与古典解决者密切配合,但他们无法在实际规模上将所学到的政策推广到更大的实例。这项工作展示了一个终端到终端的神经组合优化管道,它汇集了最近的一些论文,以便找出诱导偏差、模型结构和学习算法,促进普遍化到比培训中看到的更大的例子。我们控制的实验为这种零光谱化提供了第一次原则性调查,揭示出超出培训数据外的外推法需要重新思考神经组合优化管道,从网络层到学习范式到评估协议。此外,我们分析了通过管道透镜为路由问题进行深度学习的最新进展,并为刺激未来研究提供了新的方向。