Gradient methods have become mainstream techniques for Bi-Level Optimization (BLO) in learning and vision fields. The validity of existing works heavily relies on solving a series of approximation subproblems with extraordinarily high accuracy. Unfortunately, to achieve the approximation accuracy requires executing a large quantity of time-consuming iterations and computational burden is naturally caused. This paper is thus devoted to address this critical computational issue. In particular, we propose a single-level formulation to uniformly understand existing explicit and implicit Gradient-based BLOs (GBLOs). This together with our designed counter-example can clearly illustrate the fundamental numerical and theoretical issues of GBLOs and their naive accelerations. By introducing the dual multipliers as a new variable, we then establish Bilevel Alternating Gradient with Dual Correction (BAGDC), a general framework, which significantly accelerates different categories of existing methods by taking specific settings. A striking feature of our convergence result is that, compared to those original unaccelerated GBLO versions, the fast BAGDC admits a unified non-asymptotic convergence theory towards stationarity. A variety of numerical experiments have also been conducted to demonstrate the superiority of the proposed algorithmic framework.
翻译:渐进方法已成为学习和视觉领域双优化的主流技术(BLO),现有工程的有效性在很大程度上取决于如何解决一系列近似子问题,其精度极高。不幸的是,为了实现近近精度,自然需要执行大量耗时的迭代和计算负担。因此,本文件专门用来处理这个关键的计算问题。特别是,我们提议一个单一层次的公式,以统一理解现有的明确和隐含的“基于渐进的”双轨(GBLOs),这与我们设计的反例一道,可以清楚地说明GBLOs的基本数字和理论问题及其天性加速。通过采用双重乘数作为新的变量,我们随后建立双向变数,并用双重校正(BAGDC)这个总框架,按照具体环境大大加快了现有方法的不同类别。我们趋同结果的一个突出特征是,与原有的未加加速的GBLOs版本相比,快速的BAGDC承认一个统一的非被动趋同性趋同性理论,也展示了拟议中的矩阵性矩阵。