Kernel Regularized Least Squares (KRLS) is a popular method for flexibly estimating models that may have complex relationships between variables. However, its usefulness to many researchers is limited for two reasons. First, existing approaches are inflexible and do not allow KRLS to be combined with theoretically-motivated extensions such as random effects, unregularized fixed effects, or non-Gaussian outcomes. Second, estimation is extremely computationally intensive for even modestly sized datasets. Our paper addresses both concerns by introducing generalized KRLS (gKRLS). We note that KRLS can be re-formulated as a hierarchical model thereby allowing easy inference and modular model construction where KRLS can be used alongside random effects, splines, and unregularized fixed effects. Computationally, we also implement random sketching to dramatically accelerate estimation while incurring a limited penalty in estimation quality. We demonstrate that gKRLS can be fit on datasets with tens of thousands of observations in under one minute. Further, state-of-the-art techniques that require fitting the model over a dozen times (e.g. meta-learners) can be estimated quickly.
翻译:内核常规最小方(KRLS)是灵活估算模型的一种常用方法,这些模型在变量之间可能存在复杂的关系。然而,对于许多研究人员来说,其用处有限,原因有两个。首先,现有方法不灵活,不允许KRLS与随机效应、非常规固定效应或非高加索结果等具有理论动机的扩展相结合。第二,对即使规模不大的数据集,估算也是极具计算密集性的。我们的文件通过引入通用的KRLS(gKRLS)来处理这两个关切。我们注意到,KRLS可以重新形成一个等级模型,从而便于推断和模块模型构建,使KRLS与随机效应、螺纹和非正规固定效应一起使用。计算,我们还随机绘制草图,以大大加快估算,同时在估计质量方面受到有限的惩罚。我们证明,GKRLS可以在一分钟内对数据集进行数以万计的观测。此外,需要将模型安装十几倍时间(例如快速估算元中)的状态型技术。