Screening and working set techniques are important approaches to reducing the size of an optimization problem. They have been widely used in accelerating first-order methods for solving large-scale sparse learning problems. In this paper, we develop a new screening method called Newton screening (NS) which is a generalized Newton method with a built-in screening mechanism. We derive an equivalent KKT system for the Lasso and utilize a generalized Newton method to solve the KKT equations. Based on this KKT system, a built-in working set with a relatively small size is first determined using the sum of primal and dual variables generated from the previous iteration, then the primal variable is updated by solving a least-squares problem on the working set and the dual variable updated based on a closed-form expression. Moreover, we consider a sequential version of Newton screening (SNS) with a warm-start strategy. We show that NS possesses an optimal convergence property in the sense that it achieves one-step local convergence. Under certain regularity conditions on the feature matrix, we show that SNS hits a solution with the same signs as the underlying true target and achieves a sharp estimation error bound with high probability. Simulation studies and real data analysis support our theoretical results and demonstrate that SNS is faster and more accurate than several state-of-the-art methods in our comparative studies.
翻译:筛选和工作集技术是减少优化问题大小的重要方法。它们已广泛用于加速用于解决大规模稀疏学习问题的一阶方法。本文提出了一种新的筛选方法,称为牛顿筛选法(NS),它是一种具有内置筛选机制的广义牛顿方法。我们推导了Lasso的等效KKT系统,并利用广义牛顿方法来解决KKT方程。基于这个KKT系统,首先使用前一次迭代生成的原始和对偶变量之和确定了一个内置工作集,然后通过在工作集上求解最小二乘问题来更新主变量,并根据一个闭合形式表达式更新对偶变量。此外,我们考虑一种带有热启动策略的顺序版本的牛顿筛选(SNS)。我们表明NS具有最优收敛性质,即以一步局部收敛。在特征矩阵上的某些正则条件下,我们表明SNS与底层真实目标具有相同的符号,并以高概率实现尖锐的估计误差界限。仿真研究和实际数据分析支持我们的理论结果,并证明SNS比我们的比较研究中的几种最新方法更快、更准确。