Wasserstein distributionally robust optimization (DRO) has found success in operations research and machine learning applications as a powerful means to obtain solutions with favourable out-of-sample performances. Two compelling explanations for the success are the generalization bounds derived from Wasserstein DRO and the equivalency between Wasserstein DRO and the regularization scheme commonly applied in machine learning. Existing results on generalization bounds and the equivalency to regularization are largely limited to the setting where the Wasserstein ball is of a certain type and the decision criterion takes certain forms of an expected function. In this paper, we show that by focusing on Wasserstein DRO problems with affine decision rules, it is possible to obtain generalization bounds and the equivalency to regularization in a significantly broader setting where the Wasserstein ball can be of a general type and the decision criterion can be a general measure of risk, i.e., nonlinear in distributions. This allows for accommodating many important classification, regression, and risk minimization applications that have not been addressed to date using Wasserstein DRO. Our results are strong in that the generalization bounds do not suffer from the curse of dimensionality and the equivalency to regularization is exact. As a byproduct, our regularization results broaden considerably the class of Wasserstein DRO models that can be solved efficiently via regularization formulations.
翻译:瓦森斯坦分配强力优化(DRO)发现,在操作研究和机器学习应用方面取得成功,是获得有利外表性表现解决方案的有力手段。成功有两个令人信服的解释是瓦森斯坦DRO的概括性界限和瓦森斯坦DRO与通常用于机器学习的正规化机制之间的等同关系。关于一般化约束和对准正规化的现有结果主要限于瓦森斯坦球属于某类,而决定标准则采用某种预期功能的形式。在本文中,我们通过注重瓦森斯坦DRO的问题,以草率决定规则为主,就有可能获得普遍化界限和在相当宽得多的环境中实现正规化的等同性。瓦森斯坦DRO的常规化标准可以是一般风险的一般衡量标准,即非线性分布。这样可以容纳许多重要的分类、回归和尽量减少风险的应用,而这些应用尚未使用瓦森斯坦DRO的办法来加以解决。我们的结果是强有力的,在通过一般化的正规化模式逐步扩大,这种稳定化的稳定性是不受到普遍性约束的。