In data-driven stochastic optimization, model parameters of the underlying distribution need to be estimated from data in addition to the optimization task. Recent literature suggests the integration of the estimation and optimization processes, by selecting model parameters that lead to the best empirical objective performance. Such an integrated approach can be readily shown to outperform simple ``estimate then optimize" when the model is misspecified. In this paper, we argue that when the model class is rich enough to cover the ground truth, the performance ordering between the two approaches is reversed for nonlinear problems in a strong sense. Simple ``estimate then optimize" outperforms the integrated approach in terms of stochastic dominance of the asymptotic optimality gap, i,e, the mean, all other moments, and the entire asymptotic distribution of the optimality gap is always better. Analogous results also hold under constrained settings and when contextual features are available. We also provide experimental findings to support our theory.
翻译:在数据驱动的随机优化中,除了优化任务之外,还需要从数据中估计基础分布的模型参数。最近的文献建议通过选择导致最佳经验目标性能的模型参数来整合估计和优化过程。当模型被错误地指定时,这样的一种综合方法可以很容易地表现出优于简单的“先估计然后再优化”的方法。在本文中,我们认为,当模型类足够丰富以覆盖基本事实时,非线性问题中两种方法之间的性能排序以强烈的方式发生了反转。简单的“先估计然后再优化”方法在渐近优化差距的随机优势方面优于综合方法,即均值,所有其他时刻和整个渐近分布的优化差距总是更好。类似的结果在约束设置和当上下文特征可用时也成立。我们还提供实验结果来支持我们的理论。