Stochastic variance reduction has proven effective at accelerating first-order algorithms for solving convex finite-sum optimization tasks such as empirical risk minimization. Incorporating additional second-order information has proven helpful in further improving the performance of these first-order methods. However, comparatively little is known about the benefits of using variance reduction to accelerate popular stochastic second-order methods such as Subsampled Newton. To address this, we propose Stochastic Variance-Reduced Newton (SVRN), a finite-sum minimization algorithm which enjoys all the benefits of second-order methods: simple unit step size, easily parallelizable large-batch operations, and fast local convergence, while at the same time taking advantage of variance reduction to achieve improved convergence rates (per data pass) for smooth and strongly convex problems. We show that SVRN can accelerate many stochastic second-order methods (such as Subsampled Newton) as well as iterative least squares solvers (such as Iterative Hessian Sketch), and it compares favorably to popular first-order methods with variance reduction.
翻译:事实证明,减少骨质差异在加快一阶算法以解决诸如尽量减少实证风险等有限和优化任务方面是行之有效的。纳入额外的二阶信息有助于进一步改善这些第一阶方法的绩效。然而,相对而言,对于利用差异减少来加快诸如Subsampled Newton等流行的随机第二阶方法的好处了解甚少。为了解决这一问题,我们提议采用一个可享受第二阶方法所有好处的有限和最小化算法(SVRN),即简单单位步法大小、容易平行的大批作业和快速本地趋同,同时利用差异减少来提高平稳和强烈交错问题的趋同率(每条数据流)。我们表明,SVRN可以加速许多二阶方法(例如子标注牛顿)以及迭接式最低方位解决方案(如Iterative Hesian Scet),而且与流行的第一阶方法相比,可减少差异。