Recently, Transformer has become a prevailing deep architecture for solving vehicle routing problems (VRPs). However, it is less effective in learning improvement models for VRP because its positional encoding (PE) method is not suitable in representing VRP solutions. This paper presents a novel Dual-Aspect Collaborative Transformer (DACT) to learn embeddings for the node and positional features separately, instead of fusing them together as done in existing ones, so as to avoid potential noises and incompatible correlations. Moreover, the positional features are embedded through a novel cyclic positional encoding (CPE) method to allow Transformer to effectively capture the circularity and symmetry of VRP solutions (i.e., cyclic sequences). We train DACT using Proximal Policy Optimization and design a curriculum learning strategy for better sample efficiency. We apply DACT to solve the traveling salesman problem (TSP) and capacitated vehicle routing problem (CVRP). Results show that our DACT outperforms existing Transformer based improvement models, and exhibits much better generalization performance across different problem sizes on synthetic and benchmark instances, respectively.
翻译:最近,变异器已成为解决车辆路由问题的主流深层建筑(VRPs),但在学习VRP改进模型方面效果较差,因为其定位编码方法不适合代表VRP解决方案。本文介绍了一个新的双光合作变异器(DACT),以分别学习节点和定位特性的嵌入,而不是像现有模式那样把它们拼在一起,以避免潜在的噪音和不相容的关联。此外,定位特征通过一种全新的周期定位编码(CPE)方法嵌入,使变异器能够有效捕捉VRP解决方案(即周期序列)的循环性和对称性。我们用Proximal政策优化来培训DACT,并设计课程学习战略,以提高样本效率。我们应用DACT来解决旅行销售员问题(TSP)和电容器车辆路由问题(CVRP)。结果显示,我们的DACT系统比现有的变异器改进模型要优得多,并显示不同问题规模、合成基准和不同例子的通用性业绩要好得多。