We give sublinear-time approximation algorithms for some optimization problems arising in machine learning, such as training linear classifiers and finding minimum enclosing balls. Our algorithms can be extended to some kernelized versions of these problems, such as SVDD, hard margin SVM, and L2-SVM, for which sublinear-time algorithms were not known before. These new algorithms use a combination of a novel sampling techniques and a new multiplicative update algorithm. We give lower bounds which show the running times of many of our algorithms to be nearly best possible in the unit-cost RAM model. We also give implementations of our algorithms in the semi-streaming setting, obtaining the first low pass polylogarithmic space and sublinear time algorithms achieving arbitrary approximation factor.
翻译:我们为机器学习中出现的一些优化问题提供亚线性时间近似算法,例如培训线性分类员和找到最小附加球。我们的算法可以推广到这些问题的某些内脏化版本,如SVDD、硬边SVM和L2-SVM,而对于这些问题,以前还不知道有亚线性亚线性算法。这些新算法使用一种新型取样技术和新的多样性更新算法的组合。我们给出了较低的界限,显示我们许多算法运行时间在单位成本的RAM模型中几乎是最佳的。我们还在半流环境中应用了我们的算法,获得了第一个低通过多面空间和亚线性时间算法,实现了任意近似系数。