Meta-learning or learning to learn is a popular approach for learning new tasks with limited data (i.e., few-shot learning) by leveraging the commonalities among different tasks. However, meta-learned models can perform poorly when context data is limited, or when data is drawn from an out-of-distribution (OoD) task. Especially in safety-critical settings, this necessitates an uncertainty-aware approach to meta-learning. In addition, the often multimodal nature of task distributions can pose unique challenges to meta-learning methods. In this work, we present UnLiMiTD (uncertainty-aware meta-learning for multimodal task distributions), a novel method for meta-learning that (1) makes probabilistic predictions on in-distribution tasks efficiently, (2) is capable of detecting OoD context data at test time, and (3) performs on heterogeneous, multimodal task distributions. To achieve this goal, we take a probabilistic perspective and train a parametric, tuneable distribution over tasks on the meta-dataset. We construct this distribution by performing Bayesian inference on a linearized neural network, leveraging Gaussian process theory. We demonstrate that UnLiMiTD's predictions compare favorably to, and outperform in most cases, the standard baselines, especially in the low-data regime. Furthermore, we show that UnLiMiTD is effective in detecting data from OoD tasks. Finally, we confirm that both of these findings continue to hold in the multimodal task-distribution setting.
翻译:元学习或学习是利用不同任务之间的共同点,利用有限数据学习新任务(即少见学习)的流行方法;然而,当背景数据有限时,或当数据来自分配外(OoD)任务时,元学习模式可能效果不佳,特别是当数据来自分配外(OoD)任务时;特别是在安全危急的情况下,这需要对元学习采取有不确定性和觉悟的方法;此外,任务分配常常是多式的,对元学习方法构成独特的挑战。在这项工作中,我们介绍了UnLiMITD(为多式联运任务分配进行不可靠的元学习),这是一种新颖的元学习方法,它(1) 对分配中的任务作出概率预测,(2) 在测试时能够检测OOD背景数据,(3) 执行多式、多式的任务分配。为了实现这一目标,我们采取一种稳定的观点,对元数据集的任务进行一种可计量的、可调和可调的分布方法。我们通过对在线的线性内线性线性网络进行推论来构建这种分布,利用高层次和最精确的模型,我们用最精确的模型来评估数据定义。