Although application examples of multilevel optimization have already been discussed since the '90s, the development of solution methods was almost limited to bilevel cases due to the difficulty of the problem. In recent years, in machine learning, Franceschi et al. have proposed a method for solving bilevel optimization problems by replacing their lower-level problems with the $T$ steepest descent update equations with some prechosen iteration number $T$. In this paper, we have developed a gradient-based algorithm for multilevel optimization with $n$ levels based on their idea and proved that our reformulation with $n T$ variables asymptotically converges to the original multilevel problem. As far as we know, this is one of the first algorithms with some theoretical guarantee for multilevel optimization. Numerical experiments show that a trilevel hyperparameter learning model considering data poisoning produces more stable prediction results than an existing bilevel hyperparameter learning model in noisy data settings.
翻译:尽管自90年代以来已经讨论了多层次优化的应用实例,但解决方案方法的制定几乎局限于双级案例,原因是问题的困难。近年来,在机器学习中,Franceschi等人提出了一种解决双级优化问题的方法,即用美元最深层的下层更新方程替换其较低层次的问题,并预选代用美元。在本文中,我们根据他们的想法制定了一种基于梯度的多级优化算法,以美元为基础,以多级优化为单位,并证明我们用n T$ 变量重新拟订的零位变量与最初的多级问题基本一致。据我们所知,这是第一种具有多级优化理论保证的理论性算法。数字实验表明,考虑到数据中毒的三级超参数学习模型比在杂乱数据环境中现有的双级超参数学习模型产生更稳定的预测结果。