Few-shot meta-learning methods consider the problem of learning new tasks from a small, fixed number of examples, by meta-learning across static data from a set of previous tasks. However, in many real world settings, it is more natural to view the problem as one of minimizing the total amount of supervision --- both the number of examples needed to learn a new task and the amount of data needed for meta-learning. Such a formulation can be studied in a sequential learning setting, where tasks are presented in sequence. When studying meta-learning in this online setting, a critical question arises: can meta-learning improve over the sample complexity and regret of standard empirical risk minimization methods, when considering both meta-training and adaptation together? The answer is particularly non-obvious for meta-learning algorithms with complex bi-level optimizations that may demand large amounts of meta-training data. To answer this question, we extend previous meta-learning algorithms to handle the variable-shot settings that naturally arise in sequential learning: from many-shot learning at the start, to zero-shot learning towards the end. On sequential learning problems, we find that meta-learning solves the full task set with fewer overall labels and achieves greater cumulative performance, compared to standard supervised methods. These results suggest that meta-learning is an important ingredient for building learning systems that continuously learn and improve over a sequence of problems.
翻译:少见的元学习方法考虑到从少数固定例子中学习新任务的问题,通过从先前一系列任务中的静态数据中进行元学习,从一系列固定例子中进行元学习。然而,在许多现实世界环境中,将问题看成是最大限度地减少监督总量 -- -- 既包括学习新任务所需的实例数量,也包括元学习所需的数据数量。这种提法可以在顺序学习环境中研究,任务按顺序排列。在研究这一在线环境中的元学习时,出现一个关键问题:在考虑元培训和适应的同时,对标准实验风险最小化方法的抽样复杂性和遗憾进行元学习能否得到改善?答案特别自然的是,对于具有复杂双级优化要求大量元培训数据的元学习算法的元学习算法而言,非明显。为了回答这个问题,我们扩展了以前的元学习算法,以便处理在顺序学习过程中自然产生的变数环境:从开始的多发式学习开始,到最终的零发式学习。关于连续学习问题,我们发现,在考虑元学习问题时,元学习可以解决全部任务算法,而总体的升级后,将逐渐地学习,比较重要的方法是逐步学习。