Although application examples of multilevel optimization have already been discussed since the 1990s, the development of solution methods was almost limited to bilevel cases due to the difficulty of the problem. In recent years, in machine learning, Franceschi et al. have proposed a method for solving bilevel optimization problems by replacing their lower-level problems with the $T$ steepest descent update equations with some prechosen iteration number $T$. In this paper, we have developed a gradient-based algorithm for multilevel optimization with $n$ levels based on their idea and proved that our reformulation asymptotically converges to the original multilevel problem. As far as we know, this is one of the first algorithms with some theoretical guarantee for multilevel optimization. Numerical experiments show that a trilevel hyperparameter learning model considering data poisoning produces more stable prediction results than an existing bilevel hyperparameter learning model in noisy data settings.
翻译:虽然自1990年代以来已经讨论了多层次优化的应用实例,但由于问题的困难,解决方案方法的制定几乎局限于双级案例;近年来,在机器学习中,Franceschi等人提出了一种解决双级优化问题的方法,用美元最深层的下降更新方程式取代其较低层次的问题,并预选了一定数额的复制费用。在本文中,我们开发了一种基于梯度的多级优化算法,根据其想法,以美元为水平,并证明我们重新制定的方法与最初的多层次问题不时相联。据我们所知,这是第一个具有多层次优化理论保证的理论性算法之一。数字实验表明,考虑到数据中毒的三级超参数学习模型所产生的预测结果比在噪音数据环境中现有的双级超参数学习模型更为稳定。