As opaque predictive models increasingly impact many areas of modern life, interest in quantifying the importance of a given input variable for making a specific prediction has grown. Recently, there has been a proliferation of model-agnostic methods to measure variable importance (VI) that analyze the difference in predictive power between a full model trained on all variables and a reduced model that excludes the variable(s) of interest. A bottleneck common to these methods is the estimation of the reduced model for each variable (or subset of variables), which is an expensive process that often does not come with theoretical guarantees. In this work, we propose a fast and flexible method for approximating the reduced model with important inferential guarantees. We replace the need for fully retraining a wide neural network by a linearization initialized at the full model parameters. By adding a ridge-like penalty to make the problem convex, we prove that when the ridge penalty parameter is sufficiently large, our method estimates the variable importance measure with an error rate of $O(\frac{1}{\sqrt{n}})$ where $n$ is the number of training samples. We also show that our estimator is asymptotically normal, enabling us to provide confidence bounds for the VI estimates. We demonstrate through simulations that our method is fast and accurate under several data-generating regimes, and we demonstrate its real-world applicability on a seasonal climate forecasting example.
翻译:由于不透明的预测模型日益影响现代生活的许多领域,人们越来越关心量化某一投入变量对作出具体预测的重要性。最近,模型-不可知性方法的激增,以测量可变重要性(VI),分析所有变量所培训的完整模型与排除利益变量的减少模型之间在预测力上的差别。这些方法的一个共同瓶颈是估计每个变量(或变量子)的减少模型(或变量子),这是一个昂贵的过程,往往没有理论保证。在这项工作中,我们提出一种快速和灵活的方法,用重要的推断保证来近似降低的模型的可变性。我们用完全模型参数开始的线性化模型取代充分再培训宽度神经网络的需要。我们加一个类似于峰值的罚款来使问题变形,我们证明当峰值参数足够大时,我们的方法估计变量的重要性,其误率通常不及理论保证。我们提出一个快速和精确的周期性模型,我们用美元来显示我们正常的预测样本数量。我们通过快速的模型来展示我们的正常的模型。