Kernel methods are widespread in machine learning; however, they are limited by the quadratic complexity of the construction, application, and storage of kernel matrices. Low-rank matrix approximation algorithms are widely used to address this problem and reduce the arithmetic and storage cost. However, we observed that for some datasets with wide intra-class variability, the optimal kernel parameter for smaller classes yields a matrix that is less well approximated by low-rank methods. In this paper, we propose an efficient structured low-rank approximation method -- the Block Basis Factorization (BBF) -- and its fast construction algorithm to approximate radial basis function (RBF) kernel matrices. Our approach has linear memory cost and floating-point operations for many machine learning kernels. BBF works for a wide range of kernel bandwidth parameters and extends the domain of applicability of low-rank approximation methods significantly. Our empirical results demonstrate the stability and superiority over the state-of-art kernel approximation algorithms.
翻译:内核方法在机器学习中很普遍;然而,由于内核基质的构造、应用和储存具有四级复杂性,内核方法受到限制。低级矩阵近似算法被广泛用于解决这一问题并降低算术和储存成本。然而,我们注意到,对于一些具有广泛内部变异性的一些数据集,对于小类而言,最佳内核参数产生一个基质,而低级方法则不那么接近。在本文中,我们建议一种高效的结构性低级近似法 -- -- 块底系数(BBF) -- -- 及其快速构建算法,以近似辐射基基函数(RBF)内核矩阵。我们的方法对许多机器学习内核具有线性内存成本和浮点操作功能。BFF为一系列广泛的内核带带带参数工作,并大大扩展了低级近核方法的适用范围。我们的经验结果显示,相对于国家内核近核接近算法来说,其稳定性和优越性。