Optimization problems are ubiquitous in our societies and are present in almost every segment of the economy. Most of these optimization problems are NP-hard and computationally demanding, often requiring approximate solutions for large-scale instances. Machine learning frameworks that learn to approximate solutions to such hard optimization problems are a potentially promising avenue to address these difficulties, particularly when many closely related problem instances must be solved repeatedly. Supervised learning frameworks can train a model using the outputs of pre-solved instances. However, when the outputs are themselves approximations, when the optimization problem has symmetric solutions, and/or when the solver uses randomization, solutions to closely related instances may exhibit large differences and the learning task can become inherently more difficult. This paper demonstrates this critical challenge, connects the volatility of the training data to the ability of a model to approximate it, and proposes a method for producing (exact or approximate) solutions to optimization problems that are more amenable to supervised learning tasks. The effectiveness of the method is tested on hard non-linear nonconvex and discrete combinatorial problems.
翻译:优化问题在我们社会中普遍存在,而且几乎存在于经济的每一个部分。这些优化问题大多是NP硬的和计算上要求很高的,往往需要大规模情况的近似解决办法。学会为这种硬优化问题找到近似解决办法的机械学习框架是解决这些困难的一个潜在有希望的途径,特别是在许多密切相关的问题必须反复解决的情况下。受监督的学习框架可以使用预先解决案例的产出来训练一个模型。然而,当产出本身接近时,当优化问题有对称解决办法时,和(或)当解决问题者使用随机化时,对密切相关的情况的解决办法可能会显示出很大的差异,而学习任务本身可能变得更加困难。本文展示了这一关键的挑战,将培训数据的波动性与模型的匹配能力联系起来,并提出了一种(精确或估计)优化问题的方法,以更便于监管的学习任务。该方法的有效性在硬非线性非对齐和离散的调问题上测试。