Support Vector Machine is one of the most classical approaches for classification and regression. Despite being studied for decades, obtaining practical algorithms for SVM is still an active research problem in machine learning. In this paper, we propose a new perspective for SVM via saddle point optimization. We provide an algorithm which achieves $(1-\epsilon)$-approximations with running time $\tilde{O}(nd+n\sqrt{d / \epsilon})$ for both separable (hard margin SVM) and non-separable cases ($\nu$-SVM ), where $n$ is the number of points and $d$ is the dimensionality. To the best of our knowledge, the current best algorithm for hard margin SVM achieved by Gilbert algorithm~\cite{gartner2009coresets} requires $O(nd / \epsilon )$ time. Our algorithm improves the running time by a factor of $\sqrt{d}/\sqrt{\epsilon}$. For $\nu$-SVM, besides the well known quadratic programming approach which requires $\Omega(n^2 d)$ time~\cite{joachims1998making,platt199912}, no better algorithm is known. In the paper, we provide the first nearly linear time algorithm for $\nu$-SVM. We also consider the distributed settings and provide distributed algorithms with low communication cost via saddle point optimization. Our algorithms require $\tilde{O}(k(d +\sqrt{d/\epsilon}))$ communication cost where $k$ is the number of clients, almost matching the theoretical lower bound.
翻译:支持矢量机是最经典的分类和回归方法之一。 尽管正在研究数十年, 获取 SVM 的实用算法仍然是机器学习中的一个积极的研究问题。 在本文中, 我们通过马鞍点优化为 SVM 提出了一个新的视角。 我们提供一种以运行时间$\ tilde{O}( d+n\ sqrt{d/\ epsilon} ) 实现$( 1\\\ epsilon) $( 硬差SVM ) 和不可分离案例( $\ nu$- SVM ) 的匹配方法。 我们的算法通过运行时间 $( sqrt{ svM ) 来改善运行时间 。 我们的算法以 $\ talder} 和 rent rexcial $( 美元) 。 对于我们所知的算法来说, 美元\\\\\\\\\\ raltial_ ral_ ral_ ral_ ral_ ralxxxxxxxxx_ lexxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx