As machine learning models are deployed in critical applications, it becomes important to not just provide point estimators of the model parameters (or subsequent predictions), but also quantify the uncertainty associated with estimating the model parameters via confidence sets. In the last decade, estimating or training in several machine learning models has become synonymous with running stochastic gradient algorithms. However, computing the stochastic gradients in several settings is highly expensive or even impossible at times. An important question which has thus far not been addressed sufficiently in the statistical machine learning literature is that of equipping zeroth-order stochastic gradient algorithms with practical yet rigorous inferential capabilities. Towards this, in this work, we first establish a central limit theorem for Polyak-Ruppert averaged stochastic gradient algorithm in the zeroth-order setting. We then provide online estimators of the asymptotic covariance matrix appearing in the central limit theorem, thereby providing a practical procedure for constructing asymptotically valid confidence sets (or intervals) for parameter estimation (or prediction) in the zeroth-order setting.
翻译:由于机器学习模型部署在关键应用程序中,不仅提供模型参数(或随后的预测)的点测算器,而且通过信心组合来量化与估计模型参数有关的不确定性,都变得非常重要。在过去十年中,若干机器学习模型的估算或培训已经成为运行随机梯度算法的同义词。然而,在几种情况下计算随机梯度的费用非常昂贵,有时甚至不可能。统计机器学习文献中迄今未充分处理的一个重要问题是,为零顺序随机梯度算法配备实用但严格的推论能力。为此,我们首先在零顺序设置中为聚氨-拉普特平均梯度算法设定了一个中心限值。我们然后提供位于中央界限的无干扰共变矩阵的在线估计符,从而为在零顺序设置中为参数估算(或预测)构建无源有效信任套件(或间隔)提供一个实用程序。