Online meta-learning has recently emerged as a marriage between batch meta-learning and online learning, for achieving the capability of quick adaptation on new tasks in a lifelong manner. However, most existing approaches focus on the restrictive setting where the distribution of the online tasks remains fixed with known task boundaries. In this work, we relax these assumptions and propose a novel algorithm for task-agnostic online meta-learning in non-stationary environments. More specifically, we first propose two simple but effective detection mechanisms of task switches and distribution shift based on empirical observations, which serve as a key building block for more elegant online model updates in our algorithm: the task switch detection mechanism allows reusing of the best model available for the current task at hand, and the distribution shift detection mechanism differentiates the meta model update in order to preserve the knowledge for in-distribution tasks and quickly learn the new knowledge for out-of-distribution tasks. In particular, our online meta model updates are based only on the current data, which eliminates the need of storing previous data as required in most existing methods. We further show that a sublinear task-averaged regret can be achieved for our algorithm under mild conditions. Empirical studies on three different benchmarks clearly demonstrate the significant advantage of our algorithm over related baseline approaches.
翻译:最近,在线元学习已成为分批元学习和在线学习之间的一种结合,目的是实现以终身方式快速适应新任务的能力。然而,大多数现有方法侧重于限制设置,使在线任务的分配仍然固定在已知的任务界限内。在这项工作中,我们放松这些假设,提出一个新的算法,用于在非静止环境中进行任务不可知的在线元学习。更具体地说,我们首先提出基于实证观察的两个简单而有效的任务开关检测和分配转移机制,作为我们算法中更优雅的在线模式更新的关键基石:任务交换检测机制允许重新使用目前任务的最佳模式,分配转移检测机制区分元模式更新,以便保留分配任务的知识,并迅速学习分配任务外分配任务的新知识。特别是,我们的在线元模型更新仅以现有数据为基础,从而消除了按大多数现有方法需要存储先前数据的需求。我们进一步表明,在温和的条件下,可以实现分线任务平均的遗憾分线,用于我们的算法。