Supervised learning can improve the design of state-of-the-art solvers for combinatorial problems, but labelling large numbers of combinatorial instances is often impractical due to exponential worst-case complexity. Inspired by the recent success of contrastive pre-training for images, we conduct a scientific study of the effect of augmentation design on contrastive pre-training for the Boolean satisfiability problem. While typical graph contrastive pre-training uses label-agnostic augmentations, our key insight is that many combinatorial problems have well-studied invariances, which allow for the design of label-preserving augmentations. We find that label-preserving augmentations are critical for the success of contrastive pre-training. We show that our representations are able to achieve comparable test accuracy to fully-supervised learning while using only 1% of the labels. We also demonstrate that our representations are more transferable to larger problems from unseen domains. Our code is available at https://github.com/h4duan/contrastive-sat.
翻译:受监督的学习可以改进组合问题最先进的解决方案的设计,但是由于指数性最坏情况的复杂性,给大量组合实例贴上标签往往不切实际。由于最近对图像的对比性预培训的成功,我们开展了一项科学研究,研究增强型设计对于对布利安卫星卫星可容性问题的对比性预培训的影响。典型的图表对比性前培训使用标签 -- -- 不可知性扩增,我们的关键见解是,许多组合性问题存在研究周密的变异性,从而可以设计标签保留增强。我们发现,标签保留增强对于对比性预培训的成功至关重要。我们显示,我们的表现能够达到可比的测试精度,以完全超强的学习,同时只使用1%的标签。我们还表明,我们的表达方式更容易从看不见的领域转移到更大的问题。我们的代码可以在 http://github.com/h4duan/contrastrastive-sat上查阅。