Estimation of the precision matrix (or inverse covariance matrix) is of great importance in statistical data analysis. However, as the number of parameters scales quadratically with the dimension p, computation becomes very challenging when p is large. In this paper, we propose an adaptive sieving reduction algorithm to generate a solution path for the estimation of precision matrices under the $\ell_1$ penalized D-trace loss, with each subproblem being solved by a second-order algorithm. In each iteration of our algorithm, we are able to greatly reduce the number of variables in the problem based on the Karush-Kuhn-Tucker (KKT) conditions and the sparse structure of the estimated precision matrix in the previous iteration. As a result, our algorithm is capable of handling datasets with very high dimensions that may go beyond the capacity of the existing methods. Moreover, for the sub-problem in each iteration, other than solving the primal problem directly, we develop a semismooth Newton augmented Lagrangian algorithm with global linear convergence on the dual problem to improve the efficiency. Theoretical properties of our proposed algorithm have been established. In particular, we show that the convergence rate of our algorithm is asymptotically superlinear. The high efficiency and promising performance of our algorithm are illustrated via extensive simulation studies and real data applications, with comparison to several state-of-the-art solvers.
翻译:精确矩阵(或反常变矩阵)的估算在统计数据分析中非常重要。然而,随着参数数量与维度的量度比例的大小,计算在p大的时候变得非常具有挑战性。在本文中,我们提出一个适应性筛选递减算法,以产生一种解决方案路径,用于在受罚Dtrax损失的$_1美元下估算精确矩阵,每个子问题都由二级算法解决。在我们的算法的每一次迭代中,我们都能大大减少问题变量的数量,其依据是Karush-Kuhn-Tucker(KKTT)的条件和先前迭代中估计精确矩阵的稀薄结构。因此,我们提出的算法能够以非常高的尺寸处理可能超出现有方法能力的数据集。此外,对于每种代号中的子问题,除了直接解决原始问题之外,我们还开发了半斯mooth Newton 增强Lagrangian 算法,在双重问题上全球线性趋同性趋同,从而提高我们所拟的高级运算法的效率。