Sparsity regularized loss minimization problems play an important role in various fields including machine learning, data mining, and modern statistics. Proximal gradient descent method and coordinate descent method are the most popular approaches to solving the minimization problem. Although existing methods can achieve implicit model identification, aka support set identification, in a finite number of iterations, these methods still suffer from huge computational costs and memory burdens in high-dimensional scenarios. The reason is that the support set identification in these methods is implicit and thus cannot explicitly identify the low-complexity structure in practice, namely, they cannot discard useless coefficients of the associated features to achieve algorithmic acceleration via dimension reduction. To address this challenge, we propose a novel accelerated doubly stochastic gradient descent (ADSGD) method for sparsity regularized loss minimization problems, which can reduce the number of block iterations by eliminating inactive coefficients during the optimization process and eventually achieve faster explicit model identification and improve the algorithm efficiency. Theoretically, we first prove that ADSGD can achieve a linear convergence rate and lower overall computational complexity. More importantly, we prove that ADSGD can achieve a linear rate of explicit model identification. Numerically, experimental results on benchmark datasets confirm the efficiency of our proposed method.
翻译:在各种领域,包括机器学习、数据挖掘和现代统计领域,使损失减少问题成为固定损失减少问题。准梯度梯度下降方法和协调下降方法是解决最小化问题最受欢迎的方法。虽然现有方法可以实现隐含的模型识别,但是在有限的迭代中,Aka支持成套识别方法仍然在高维度情景中面临巨大的计算成本和记忆负担。原因是这些方法中的支持性确定方法隐含着隐含,因此无法明确确定实践中的低复杂结构,即它们不能丢弃相关特征的无用的系数,通过降低维度加速算法加速。为了应对这一挑战,我们提议采用新的加速双相偏梯度梯度梯度下降方法(ADSGD),以缓解重度调整损失减少问题,这可以通过在优化过程中消除不活动系数,并最终实现更快明确的模型识别并改进算法效率。理论上,我们首先证明ADSGD能够实现线性趋加速系数,并降低总体计算复杂性。更重要的是,我们证明ADSGDGD能够实现明确模型确定结果的直线性标准。