Kernel methods are fundamental in machine learning, and faster algorithms for kernel approximation provide direct speedups for many core tasks in machine learning. The polynomial kernel is especially important as other kernels can often be approximated by the polynomial kernel via a Taylor series expansion. Recent techniques in oblivious sketching reduce the dependence in the running time on the degree $q$ of the polynomial kernel from exponential to polynomial, which is useful for the Gaussian kernel, for which $q$ can be chosen to be polylogarithmic. However, for more slowly growing kernels, such as the neural tangent and arc-cosine kernels, $q$ needs to be polynomial, and previous work incurs a polynomial factor slowdown in the running time. We give a new oblivious sketch which greatly improves upon this running time, by removing the dependence on $q$ in the leading order term. Combined with a novel sampling scheme, we give the fastest algorithms for approximating a large family of slow-growing kernels.
翻译:内核法是机器学习的基础,而内核近似法的快速算法为机器学习中的许多核心任务提供了直接的加速。 多元内核特别重要, 因为其他内核往往可以通过泰勒系列扩展被多元内核所近似。 隐性草图的最近技术减少了从指数到多元内核的运行时间对从指数到多元内核的水平的依赖性。 这对高斯内核是有用的, 高斯内核可以选择美元为多元性。 但是, 对于较缓慢增长的内核, 如神经色和弧- 圆内核等, 美元需要是多元的, 先前的工作在运行时会减慢。 我们给出一种新的隐性草图, 这会在这种运行时大大改善, 在最初的顺序期, 消除对美元的依赖性。 再加上新的取样方案, 我们给一个较快速的算法, 来适应一个大型的慢层家庭。