In this paper, we study the predict-then-optimize problem where the output of a machine learning prediction task is used as the input of some downstream optimization problem, say, the objective coefficient vector of a linear program. The problem is also known as predictive analytics or contextual linear programming. The existing approaches largely suffer from either (i) optimization intractability (a non-convex objective function)/statistical inefficiency (a suboptimal generalization bound) or (ii) requiring strong condition(s) such as no constraint or loss calibration. We develop a new approach to the problem called \textit{maximum optimality margin} which designs the machine learning loss function by the optimality condition of the downstream optimization. The max-margin formulation enjoys both computational efficiency and good theoretical properties for the learning procedure. More importantly, our new approach only needs the observations of the optimal solution in the training data rather than the objective function, which makes it a new and natural approach to the inverse linear programming problem under both contextual and context-free settings; we also analyze the proposed method under both offline and online settings, and demonstrate its performance using numerical experiments.
翻译:在本文中,我们研究了将机器学习预测任务的产出用作某些下游优化问题投入的预测-最佳化问题,例如线性程序的目标系数矢量。这个问题也被称为预测分析或上下文线性编程。现有方法主要有以下两种问题:(一) 优化可吸引性(非凝固目标功能)/统计效率低下(非优化一般化约束)/统计性效率低下(非优化一般化约束)或(二) 要求条件强(如没有约束或损失校准)的问题。我们开发了一种名为\text{最高最佳差值}的新办法,根据下游优化条件设计机器学习损失函数。最大差值公式既具有计算效率,又具有学习程序的良好理论属性。更重要的是,我们的新办法只需要对培训数据的最佳解决方案而不是客观功能进行观察,从而在背景和背景上均无环境下对反线性编程问题采取新的自然办法;我们还分析了在离线和在线环境中拟议的方法,并利用数字实验来展示其表现。