We develop an exact coordinate descent algorithm for high-dimensional regularized Huber regression. In contrast to composite gradient descent methods, our algorithm fully exploits the advantages of coordinate descent when the underlying model is sparse. Moreover, unlike existing second-order approximation methods previously introduced in the literature, it remains effective even when the Hessian becomes ill-conditioned due to high correlations among covariates drawn from heavy-tailed distributions. The key idea is that, for each coordinate, marginal increments arise only from inlier observations, while the derivatives remain monotonically increasing over a grid constructed from the partial residuals. Building on conventional coordinate descent strategies, we further propose variable screening rules that selectively determine which variables to update at each iteration, thereby accelerating convergence. To the best of our knowledge, this is the first work to develop a first-order coordinate descent algorithm for penalized Huber loss minimization. We bound the nonasymptotic convergence rate of the proposed algorithm by extending arguments developed for the Lasso and formally characterize the operation of the proposed screening rule. Extensive simulation studies under heavy-tailed and highly-correlated predictors, together with a real data application, demonstrate both the practical efficiency of the method and the benefits of the computational enhancements.
翻译:本文针对高维正则化Huber回归提出了一种精确坐标下降算法。与复合梯度下降方法相比,当基础模型具有稀疏性时,本算法能充分发挥坐标下降的优势。此外,与文献中已有的二阶近似方法不同,即使协变量来自重尾分布且存在高度相关性导致海森矩阵病态时,本方法依然有效。其核心思想在于:对于每个坐标,边际增量仅来源于内点观测值,而导数在基于偏残差构建的网格上保持单调递增。在传统坐标下降策略的基础上,我们进一步提出了变量筛选规则,该规则能选择性确定每次迭代中需要更新的变量,从而加速收敛。据我们所知,这是首个针对惩罚化Huber损失最小化问题开发的一阶坐标下降算法。通过扩展针对Lasso问题构建的理论框架,我们界定了所提算法的非渐近收敛速率,并形式化描述了所提筛选规则的运行机制。在重尾分布和高度相关预测变量场景下的大量仿真研究,结合实际数据应用,共同验证了本方法的实际效率与计算增强带来的优势。