Recent advances in the field of meta-learning have tackled domains consisting of large numbers of small ("few-shot") supervised learning tasks. Meta-learning algorithms must be able to rapidly adapt to any individual few-shot task, fitting to a small support set within a task and using it to predict the labels of the task's query set. This problem setting can be extended to the Bayesian context, wherein rather than predicting a single label for each query data point, a model predicts a distribution of labels capturing its uncertainty. Successful methods in this domain include Bayesian ensembling of MAML-based models, Bayesian neural networks, and Gaussian processes with learned deep kernel and mean functions. While Gaussian processes have a robust Bayesian interpretation in the meta-learning context, they do not naturally model non-Gaussian predictive posteriors for expressing uncertainty. In this paper, we design a theoretically principled method, VMGP, extending Gaussian-process-based meta-learning to allow for high-quality, arbitrary non-Gaussian uncertainty predictions. On benchmark environments with complex non-smooth or discontinuous structure, we find our VMGP method performs significantly better than existing Bayesian meta-learning baselines.
翻译:元学习领域最近的进展已经解决了由大量小型(“few-shot”)监管的学习任务组成的领域。元学习算法必须能够迅速适应任何单个的微小任务,适应任务范围内的小型支助,并用来预测任务查询组的标签。这个问题设置可以扩展到巴伊西亚环境,而不必预测每个查询数据点的单一标签,一个模型预测标签的分布,捕捉其不确定性。这一领域的成功方法包括:贝伊西亚人集合以MAML为基础的模型、贝伊西亚神经网络和高萨进程,具有深层内核和中值功能。虽然高萨进程在元学习环境中有强大的贝叶西亚人解释,但它们并不自然地模拟非加萨预测的预测海象来表达不确定性。在本文中,我们设计了一个理论原则方法,即VMGP,扩大高斯-加亚进程元学习的范围,以允许高质、任意的非加萨尼神经网络,以及高萨进程进程进程进程,以及具有深层次和中值功能的过程。虽然高萨进程在元学习环境中有较复杂的基准环境,但更精确地进行不精确的研究。