Online gradient methods, like the online gradient algorithm (OGA), often depend on tuning parameters that are difficult to set in practice. We consider an online meta-learning scenario, and we propose a meta-strategy to learn these parameters from past tasks. Our strategy is based on the minimization of a regret bound. It allows to learn the initialization and the step size in OGA with guarantees. We provide a regret analysis of the strategy in the case of convex losses. It suggests that, when there are parameters $\theta_1,\dots,\theta_T$ solving well tasks $1,\dots,T$ respectively and that are close enough one to each other, our strategy indeed improves on learning each task in isolation.
翻译:在线梯度方法,如在线梯度算法(OGA),往往取决于在实践中难以设定的调试参数。我们考虑了在线元学习方案,我们提出了从过去的任务中学习这些参数的元战略。我们的战略的基础是尽量减少遗憾。它允许在OGA中以担保的方式学习初始化和步骤大小。我们提供了在Convex损失情况下对战略的遗憾分析。它表明,当参数为$\theta__1,\dots,\theta_T$分别解决1,\dots,T$等好任务时,我们的战略确实在孤立地学习每项任务方面有所改进。