Learning to Optimize (L2O) has drawn increasing attention as it often remarkably accelerates the optimization procedure of complex tasks by ``overfitting" specific task type, leading to enhanced performance compared to analytical optimizers. Generally, L2O develops a parameterized optimization method (i.e., ``optimizer") by learning from solving sample problems. This data-driven procedure yields L2O that can efficiently solve problems similar to those seen in training, that is, drawn from the same ``task distribution". However, such learned optimizers often struggle when new test problems come with a substantially deviation from the training task distribution. This paper investigates a potential solution to this open challenge, by meta-training an L2O optimizer that can perform fast test-time self-adaptation to an out-of-distribution task, in only a few steps. We theoretically characterize the generalization of L2O, and further show that our proposed framework (termed as M-L2O) provably facilitates rapid task adaptation by locating well-adapted initial points for the optimizer weight. Empirical observations on several classic tasks like LASSO and Quadratic, demonstrate that M-L2O converges significantly faster than vanilla L2O with only $5$ steps of adaptation, echoing our theoretical results. Codes are available in https://github.com/VITA-Group/M-L2O.
翻译:学习优化( L2O) 已经引起越来越多的注意, 因为它常常通过“ 更新” 特定任务类型, 大大加快复杂任务的最佳程序, 导致与分析优化者相比, 提高绩效。 一般来说, L2O 开发了一个参数化优化方法( 即“ 优化” ), 学习解决抽样问题。 这个数据驱动程序产生L2O, 能够有效解决与培训中看到的问题相似的问题, 即来自同一“ 任务分配 ” 。 然而, 这些学习过的优化者往往在新的测试问题出现时, 与培训任务分配大相径庭, 从而大大偏离了培训任务分配。 本文调查了这一公开挑战的潜在解决办法,通过对L2O 优化进行元化培训, 仅以几步方式对分配任务进行快速测试- 自我适应。 我们理论上认为L2O 的通用框架( 称为 M- L2O), 并且进一步表明,我们拟议的框架( 以M- L2 GroupL) 可以促进任务快速适应,, 确定优化的初始点, 优化重量。</s>