Regularized kernel-based methods such as support vector machines (SVMs) typically depend on the underlying probability measure $\mathrm{P}$ (respectively an empirical measure $\mathrm{D}_n$ in applications) as well as on the regularization parameter $\lambda$ and the kernel $k$. Whereas classical statistical robustness only considers the effect of small perturbations in $\mathrm{P}$, the present paper investigates the influence of simultaneous slight variations in the whole triple $(\mathrm{P},\lambda,k)$, respectively $(\mathrm{D}_n,\lambda_n,k)$, on the resulting predictor. Existing results from the literature are considerably generalized and improved. In order to also make them applicable to big data, where regular SVMs suffer from their super-linear computational requirements, we show how our results can be transferred to the context of localized learning. Here, the effect of slight variations in the applied regionalization, which might for example stem from changes in $\mathrm{P}$ respectively $\mathrm{D}_n$, is considered as well.
翻译:支持矢量机(SVMS)等基于内核的常规方法通常取决于基本概率度量 $\ mathrm{P}$(在应用中,分别是实验性度量$\ mathrm{D ⁇ n$),以及正规化参数$\ lambda$和内核$k$。典型的统计稳健性仅考虑小扰动以$\ mathrm{P}美元计算的影响,而本文则调查整个三美元( mathhrm{P},\lambda,k) 美元(分别是(mathrm{D ⁇ n,\lambda_n,k)美元)同时略有变化的影响,分别取决于所生成的预测值。文献的现有结果相当普遍并得到了改进。为了将其适用于大数据,如果普通的SVMSMs因超线计算要求而受到影响,则我们展示了如何将我们的结果转移到本地化学习的背景。在这里,应用的区域化中略有变化的影响,例如可能来自$\\\\\\\\\\\\\\\ $ $ $ $ $(分别)