We analyze a general class of bilevel problems, in which the upper-level problem consists in the minimization of a smooth objective function and the lower-level problem is to find the fixed point of a smooth contraction map. This type of problems include instances of meta-learning, hyperparameter optimization and data poisoning adversarial attacks. Several recent works have proposed algorithms which warm-start the lower-level problem, i.e. they use the previous lower-level approximate solution as a staring point for the lower-level solver. This warm-start procedure allows one to improve the sample complexity in both the stochastic and deterministic settings, achieving in some cases the order-wise optimal sample complexity. We show that without warm-start, it is still possible to achieve order-wise optimal and near-optimal sample complexity for the stochastic and deterministic settings, respectively. In particular, we propose a simple method which uses stochastic fixed point iterations at the lower-level and projected inexact gradient descent at the upper-level, that reaches an $\epsilon$-stationary point using $O(\epsilon^{-2})$ and $\tilde{O}(\epsilon^{-1})$ samples for the stochastic and the deterministic setting, respectively. Compared to methods using warm-start, ours is better suited for meta-learning and yields a simpler analysis that does not need to study the coupled interactions between the upper-level and lower-level iterates.
翻译:我们分析了一般的双层问题, 高层次问题在于最大限度地减少平滑客观功能, 低层次问题在于找到平稳收缩图的固定点。 这种类型的问题包括元学习、 超参数优化和数据中毒对抗性攻击。 最近的一些工程提出了暖化引发较低层次问题的算法, 即它们使用以前的低层次近似解决办法作为较低层次解决问题者的凝视点。 这个热启动程序使得人们能够改善沙沙和确定性环境的样本复杂性, 在某些情况中, 达到顺序顺畅的上层样本复杂性。 我们显示, 没有暖化启动, 仍然有可能分别实现秩序顺畅的最佳和接近最佳的样本复杂性。 特别是, 我们提出了一种简单的方法, 使用低层次和预测的异常固定点来吸引较低层次的解决问题。 高层次的梯度下降, 达到美元平定点, 在某些情况中, 使用 美元(\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\\