In the era of big data, it is desired to develop efficient machine learning algorithms to tackle massive data challenges such as storage bottleneck, algorithmic scalability, and interpretability. In this paper, we develop a novel efficient classification algorithm, called fast polynomial kernel classification (FPC), to conquer the scalability and storage challenges. Our main tools are a suitable selected feature mapping based on polynomial kernels and an alternating direction method of multipliers (ADMM) algorithm for a related non-smooth convex optimization problem. Fast learning rates as well as feasibility verifications including the efficiency of an ADMM solver with convergence guarantees and the selection of center points are established to justify theoretical behaviors of FPC. Our theoretical assertions are verified by a series of simulations and real data applications. Numerical results demonstrate that FPC significantly reduces the computational burden and storage memory of existing learning schemes such as support vector machines, Nystr\"{o}m and random feature methods, without sacrificing their generalization abilities much.
翻译:在大数据时代,人们希望开发高效的机器学习算法,以应对大量数据挑战,如存储瓶颈、算法缩放性和可解释性。在本文中,我们开发了新型高效分类算法,称为快速多球内核分类(FCC),以克服可缩放性和存储性挑战。我们的主要工具是一种基于多球内核的合适的选定地貌制图法,以及用于相关非移动锥形优化问题的倍数算法(ADMM)交替方向法(ADMM)。快速学习率以及可行性核查,包括ADMM解答器的效率,保证汇合和选择中心点,以证明FPC的理论行为是合理的。我们的理论论断通过一系列模拟和真实数据应用得到验证。数字结果表明,FPC大大降低了支持矢量机器、Nystr\"{o}m和随机特征方法等现有学习计划的计算负担和存储记忆,同时不牺牲它们的普及能力。