In sparse estimation, in which the sum of the loss function and the regularization term is minimized, methods such as the proximal gradient method and the proximal Newton method are applied. The former is slow to converge to a solution, while the latter converges quickly but is inefficient for problems such as group lasso problems. In this paper, we examine how to efficiently find a solution by finding the convergence destination of the proximal gradient method. However, the case in which the Lipschitz constant of the derivative of the loss function is unknown has not been studied theoretically, and only the Newton method has been proposed for the case in which the Lipschitz constant is known. We show that the Newton method converges when the Lipschitz constant is unknown and extend the theory. Furthermore, we propose a new quasi-Newton method that avoids Hessian calculations and improves efficiency, and we prove that it converges quickly, providing a theoretical guarantee. Finally, numerical experiments show that the proposed method can significantly improve the efficiency.
翻译:在少许估计中,损失函数和正规化术语的总和被最小化,应用了一些方法,如准度梯度法和准度牛顿法。前者慢于求得解决办法的速度,而后者很快会汇合,但对于诸如群状斜体问题之类的问题却效率不高。在本文中,我们研究如何通过找到准度梯度法的趋同目的地来有效找到解决办法。然而,对于损失函数衍生物的利普施茨常数未知的情况,没有进行理论研究,只有牛顿法才被提议为知道利普施茨常数的牛顿法。我们表明,当利普施茨常数未知时,牛顿法会汇合,并扩展理论范围。此外,我们提出了一个新的准纽顿法,避免赫西安计算并提高效率,我们证明它很快会趋同,提供了理论保证。最后,数字实验表明,拟议的方法可以大大提高效率。