We consider a class of structured fractional minimization problems, in which the numerator part of the objective is the sum of a differentiable convex function and a convex non-smooth function, while the denominator part is a convex or concave function. This problem is difficult to solve since it is non-convex. By exploiting the structure of the problem, we propose two Coordinate Descent (CD) methods for solving this problem. The proposed methods iteratively solve a one-dimensional subproblem \textit{globally}, and they are guaranteed to converge to coordinate-wise stationary points. In the case of a convex denominator, under a weak \textit{locally bounded non-convexity condition}, we prove that the optimality of coordinate-wise stationary point is stronger than that of the standard critical point and directional point. Under additional suitable conditions, CD methods converge Q-linearly to coordinate-wise stationary points. In the case of a concave denominator, we show that any critical point is a global minimum, and CD methods converge to the global minimum with a sublinear convergence rate. We demonstrate the applicability of the proposed methods to some machine learning and signal processing models. Our experiments on real-world data have shown that our method significantly and consistently outperforms existing methods in terms of accuracy.
翻译:我们考虑一类结构化的分数规划问题,在其中,目标函数的分子部分是可微的凸函数和凸非光滑函数之和,而分母部分是一个凸或凹函数。由于这个问题是非凸的,所以很难求解。通过利用问题的结构,我们提出了两种坐标下降(CD)方法来解决这个问题。这些方法通过迭代解决一个一维子问题全局最优,它们被保证收敛到分量明显的稳定点。在分母是凸函数的情况下,在弱的局部有界非凸条件下,我们证明了分量明显的稳定点的最优性强于标准的临界点和方向点。在附加合适的条件下,CD方法以Q线性收敛到分量明显的稳定点。在分母是凹函数的情况下,我们展示了任何临界点都是全局最小值,而CD方法收敛到全局最小值的速度是次线性的。我们展示了所提出的方法在一些机器学习和信号处理模型中的适用性。我们在真实数据上的实验表明,我们的方法在准确性方面显著而一致地优于现有的方法。