Decision-Focused Learning (DFL) is a paradigm for tailoring a predictive model to a downstream optimization task that uses its predictions in order to perform better on that specific task. The main technical challenge associated with DFL is that it requires being able to differentiate through the optimization problem, which is difficult due to discontinuous solutions and other challenges. Past work has largely gotten around this issue by handcrafting task-specific surrogates to the original optimization problem that provide informative gradients when differentiated through. However, the need to handcraft surrogates for each new task limits the usability of DFL. In addition, there are often no guarantees about the convexity of the resulting surrogates and, as a result, training a predictive model using them can lead to inferior local optima. In this paper, we do away with surrogates altogether and instead learn loss functions that capture task-specific information. To the best of our knowledge, ours is the first approach that entirely replaces the optimization component of decision-focused learning with a loss that is automatically learned. Our approach (a) only requires access to a black-box oracle that can solve the optimization problem and is thus generalizable, and (b) can be convex by construction and so can be easily optimized over. We evaluate our approach on three resource allocation problems from the literature and find that our approach outperforms learning without taking into account task structure in all three domains, and even hand-crafted surrogates from the literature.
翻译:与DFL相关的主要技术挑战是,它需要能够通过优化问题加以区分,而优化问题由于不连贯的解决方案和其他挑战而变得很困难。过去的工作在很大程度上围绕这个问题,通过手工制作特定任务替代器来取代最初的优化问题,而最初的优化问题在差异中提供了信息性梯度。然而,每次新任务都需要手动替代器来取代优化,这限制了DFL的可用性。此外,我们的方法往往无法保证由此产生的代孕的趋同性,因此,培训一种预测模型可以导致当地偏差。在本文中,我们完全摆脱了套装,而不是学习收集特定任务信息的损失功能。我们的知识最丰富,我们的第一个方法完全取代了以决定为重点的学习的优化部分,而一个自动学习的损失。我们的方法(a)只需要从一个黑箱或手动的域进入一个能够解决所有任务特定任务的信息配置。因此,我们可以从三个领域进入一个黑箱和手动的域,这样我们就能通过最优化的方法来解决所有任务分配问题。