In this paper, we propose a learning algorithm that enables a model to quickly exploit commonalities among related tasks from an unseen task distribution, before quickly adapting to specific tasks from that same distribution. We investigate how learning with different task distributions can first improve adaptability by meta-finetuning on related tasks before improving goal task generalization with finetuning. Synthetic regression experiments validate the intuition that learning to meta-learn improves adaptability and consecutively generalization. Experiments on more complex image classification, continual regression, and reinforcement learning tasks demonstrate that learning to meta-learn generally improves task-specific adaptation. The methodology, setup, and hypotheses in this proposal were positively evaluated by peer review before conclusive experiments were carried out.
翻译:在本文中,我们提出一种学习算法,使模型能够快速利用从无形任务分布中相关任务之间的共性,然后迅速适应从同一分配中产生的具体任务。我们调查不同任务分布的学习如何能够首先通过对相关任务进行元调整来改善适应性,然后通过微调改进目标任务的一般化。合成回归实验证实了一种直觉,即学习元学习会改善适应性和连续的概括化。关于更复杂的图像分类、持续回归和强化学习任务的实验表明,学习元学习通常会改善特定任务的适应性。本提案中的方法、设置和假设在进行结论性实验之前得到了同侪审查的积极评价。