Kernel methods are powerful learning methodologies that provide a simple way to construct nonlinear algorithms from linear ones. Despite their popularity, they suffer from poor scalability in big data scenarios. Various approximation methods, including random feature approximation, have been proposed to alleviate the problem. However, the statistical consistency of most of these approximate kernel methods is not well understood except for kernel ridge regression wherein it has been shown that the random feature approximation is not only computationally efficient but also statistically consistent with a minimax optimal rate of convergence. In this paper, we investigate the efficacy of random feature approximation in the context of kernel principal component analysis (KPCA) by studying the trade-off between computational and statistical behaviors of approximate KPCA. We show that the approximate KPCA is both computationally and statistically efficient compared to KPCA in terms of the error associated with reconstructing a kernel function based on its projection onto the corresponding eigenspaces. The analysis hinges on Bernstein-type inequalities for the operator and Hilbert-Schmidt norms of a self-adjoint Hilbert-Schmidt operator-valued U-statistics, which is of independent interest.
翻译:内核法是强大的学习方法,它提供了从线性角度构建非线性算法的简单方法。尽管广受欢迎,但它们在大数据假设中也存在可缩放性差强的问题。提出了各种近似方法,包括随机地貌近似近似法,以缓解问题。然而,除了内核脊脊回归外,大多数近似内核方法的统计一致性还没有得到很好理解,因为其中显示随机地貌近离子不仅在计算上有效,而且在统计上也符合最小最佳趋同率。在本文中,我们通过研究近似KPA的计算和统计行为之间的取舍取舍(KPCA),调查随机地貌近似于内核主部分分析(KPCA)中随机地貌特征近近近近(KPCA)的功效。我们表明,就与KPCA相关的错误而言,KPCA在计算和统计上都是有效的,因为根据对相应电子元空间的预测重建内核函数。分析以Bernstein 和Hilbert-Smidt 标准为操作员的独立利益。