The stochastic gradient descent (SGD) algorithm has been widely used in statistical estimation for large-scale data due to its computational and memory efficiency. While most existing works focus on the convergence of the objective function or the error of the obtained solution, we investigate the problem of statistical inference of true model parameters based on SGD when the population loss function is strongly convex and satisfies certain smoothness conditions. Our main contributions are two-fold. First, in the fixed dimension setup, we propose two consistent estimators of the asymptotic covariance of the average iterate from SGD: (1) a plug-in estimator, and (2) a batch-means estimator, which is computationally more efficient and only uses the iterates from SGD. Both proposed estimators allow us to construct asymptotically exact confidence intervals and hypothesis tests. Second, for high-dimensional linear regression, using a variant of the SGD algorithm, we construct a debiased estimator of each regression coefficient that is asymptotically normal. This gives a one-pass algorithm for computing both the sparse regression coefficients and confidence intervals, which is computationally attractive and applicable to online data.
翻译:由于计算和记忆效率,大规模数据统计估计广泛使用了随机梯度梯度下降算法(SGD),因为其计算和记忆效率,大量现有工作侧重于目标功能的趋同或获得的解决方案的错误,虽然大多数现有工作侧重于目标功能的趋同或获得的解决方案的错误,但我们调查了在人口损失功能强烈相向和满足一定的平滑条件下,基于SGD的真正模型参数的统计推论问题。我们的主要贡献是双重的。首先,在固定尺寸设置中,我们建议两个一致的测算标准,即SGD中平均偏差的静态变量:(1) 插入估计器;和(2) 批量平均值估测器,在计算上效率更高,只使用SGD的推算器。两个拟议的估计器都允许我们从速构建精确的信任间隔和假设测试。第二,对于高维线性回归,我们建议使用SGD算法变量,对每个正常的回归系数设置一个不偏差的估测度估测算器。这给人带来一种有吸引力的在线测算法,用来计算微的测算。