Meta-learning, or learning-to-learn, seeks to design algorithms that can utilize previous experience to rapidly learn new skills or adapt to new environments. Representation learning---a key tool for performing meta-learning---learns a data representation that can transfer knowledge across multiple tasks, which is essential in regimes where data is scarce. Despite a recent surge of interest in the practice of meta-learning, the theoretical underpinnings of meta-learning algorithms are lacking, especially in the context of learning transferable representations. In this paper, we focus on the problem of multi-task linear regression---in which multiple linear regression models share a common, low-dimensional linear representation. Here, we provide provably fast, sample-efficient algorithms to address the dual challenges of (1) learning a common set of features from multiple, related tasks, and (2) transferring this knowledge to new, unseen tasks. Both are central to the general problem of meta-learning. Finally, we complement these results by providing information-theoretic lower bounds on the sample complexity of learning these linear features.
翻译:尽管最近对元学习做法的兴趣激增,但元学习算法的理论基础仍然缺乏,特别是在学习可转移表达方式方面。在本文中,我们侧重于多任务线性回归问题,多线性回归模型在其中具有共同的、低维的线性代表形式。在这里,我们提供了可察觉的快速、样本高效的算法,以应对以下双重挑战:(1) 从多重相关任务中学习一套共同特征,(2) 将这种知识转移到新的、看不见的任务中。这两种知识都是元学习的一般性问题的核心。最后,我们通过提供有关学习这些线性特征的抽样复杂性的信息-理论较低界限来补充这些结果。