Optimal transport (OT) is a powerful geometric tool for comparing two distributions and has been employed in various machine learning applications. In this work, we propose a novel OT formulation that takes feature correlations into account while learning the transport plan between two distributions. We model the feature-feature relationship via a symmetric positive semi-definite Mahalanobis metric in the OT cost function. For a certain class of regularizers on the metric, we show that the optimization strategy can be considerably simplified by exploiting the problem structure. For high-dimensional data, we additionally propose suitable low-dimensional modeling of the Mahalanobis metric. Overall, we view the resulting optimization problem as a non-linear OT problem, which we solve using the Frank-Wolfe algorithm. Empirical results on the discriminative learning setting, such as tag prediction and multi-class classification, illustrate the good performance of our approach.
翻译:最佳运输(OT)是比较两种分布的强大几何工具,并已用于各种机器学习应用。在这项工作中,我们提出一个新的OT配方,在学习两种分布之间的运输计划时,考虑到特征相关性。我们在OT成本函数中通过对称正对正半无穷的马哈拉诺比标准来模拟特征-特性关系。对于某类衡量标准上的规范者,我们表明,利用问题结构可以大大简化优化战略。对于高维数据,我们又建议采用适当的马哈拉诺比指标的低维度模型。总的来说,我们将由此产生的优化问题视为非线性OT问题,我们用弗兰克-沃尔夫算法来解决。关于歧视性学习环境的实证结果,例如标签预测和多级分类,说明我们方法的良好表现。