By the asymptotic oracle property, non-convex penalties represented by minimax concave penalty (MCP) and smoothly clipped absolute deviation (SCAD) have attracted much attentions in high-dimensional data analysis, and have been widely used in signal processing, image restoration, matrix estimation, etc. However, in view of their non-convex and non-smooth characteristics, they are computationally challenging. Almost all existing algorithms converge locally, and the proper selection of initial values is crucial. Therefore, in actual operation, they often combine a warm-starting technique to meet the rigid requirement that the initial value must be sufficiently close to the optimal solution of the corresponding problem. In this paper, based on the DC (difference of convex functions) property of MCP and SCAD penalties, we aim to design a global two-stage algorithm for the high-dimensional least squares linear regression problems. A key idea for making the proposed algorithm to be efficient is to use the primal dual active set with continuation (PDASC) method, which is equivalent to the semi-smooth Newton (SSN) method, to solve the corresponding sub-problems. Theoretically, we not only prove the global convergence of the proposed algorithm, but also verify that the generated iterative sequence converges to a d-stationary point. In terms of computational performance, the abundant research of simulation and real data show that the algorithm in this paper is superior to the latest SSN method and the classic coordinate descent (CD) algorithm for solving non-convex penalized high-dimensional linear regression problems.
翻译:以非隐性或触角属性为代表的非隐性惩罚,由迷你马克思二次曲线罚款(MCP)和平稳剪切绝对偏差(SCAD)代表的非隐性惩罚在高维数据分析中引起了许多注意,并被广泛用于信号处理、图像恢复、矩阵估计等。然而,鉴于其非隐性和非显性特性,这些惩罚在计算上具有挑战性。几乎所有现有的算法都聚集在本地,而正确选择初始值至关重要。因此,在实际操作中,它们往往结合一种热启动技术,以满足僵硬的要求,即初始值必须足够接近相应问题的最佳解决方案。在本文中,基于DC( convex函数的差异)、 MCP 和 SCAD 刑罚的属性被广泛用于信号处理。 然而,几乎所有现有的算法都是使拟议算法具有效率的原始双向主动(PDASC) 方法,这相当于对牛顿的精度快速递增性递增性递增性(SSN) 方法, 也就是我们所拟的极性递增性排序。