The task of approximating an arbitrary convex function arises in several learning problems such as convex regression, learning with a difference of convex (DC) functions, and learning Bregman or $f$-divergences. In this paper, we develop and analyze an approach for solving a broad range of convex function learning problems that is faster than state-of-the-art approaches. Our approach is based on a 2-block ADMM method where each block can be computed in closed form. For the task of convex Lipschitz regression, we establish that our proposed algorithm converges with iteration complexity of $ O(n\sqrt{d}/\epsilon)$ for a dataset $\bm X \in \mathbb R^{n\times d}$ and $\epsilon > 0$. Combined with per-iteration computation complexity, our method converges with the rate $O(n^3 d^{1.5}/\epsilon+n^2 d^{2.5}/\epsilon+n d^3/\epsilon)$. This new rate improves the state of the art rate of $O(n^5d^2/\epsilon)$ if $d = o( n^4)$. Further we provide similar solvers for DC regression and Bregman divergence learning. Unlike previous approaches, our method is amenable to the use of GPUs. We demonstrate on regression and metric learning experiments that our approach is over 100 times faster than existing approaches on some data sets, and produces results that are comparable to state of the art.
翻译:类似任意 convex 函数的相似任务产生于若干学习问题, 如 convex 回归, 学习 convex (DC) 功能的差异, 以及学习 Bregman 或 $f美元 波动。 在本文中, 我们开发并分析一种方法, 以解决一系列广泛的 convex 函数学习问题, 其速度比最先进的方法快。 我们的方法基于一个 2 块 ADMM 方法, 每个区块都可以以封闭的形式计算 。 对于 convex 回归的任务, 我们确定我们提议的算法与 $ (n\ sqrt{d} /\\ epselon) 的折叠复杂性相匹配 。 用于一个数据集 $\\ bm X 或 in\ mathbrb Rn\ times d} 和 $\ eepslon 的递增缩缩放方法。 和 $xxxxxxx 递增缩缩缩缩缩缩缩缩 = dxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx