Meta-learning aims to extract useful inductive biases from a set of related datasets. In Bayesian meta-learning, this is typically achieved by constructing a prior distribution over neural network parameters. However, specifying families of computationally viable prior distributions over the high-dimensional neural network parameters is difficult. As a result, existing approaches resort to meta-learning restrictive diagonal Gaussian priors, severely limiting their expressiveness and performance. To circumvent these issues, we approach meta-learning through the lens of functional Bayesian neural network inference, which views the prior as a stochastic process and performs inference in the function space. Specifically, we view the meta-training tasks as samples from the data-generating process and formalize meta-learning as empirically estimating the law of this stochastic process. Our approach can seamlessly acquire and represent complex prior knowledge by meta-learning the score function of the data-generating process marginals instead of parameter space priors. In a comprehensive benchmark, we demonstrate that our method achieves state-of-the-art performance in terms of predictive accuracy and substantial improvements in the quality of uncertainty estimates.
翻译:元学习的目的是从一组相关数据集中获取有用的感化偏差。在贝耶斯元学习中,这通常是通过在神经网络参数上建立先前的分布方式来实现的。然而,很难在高维神经网络参数上确定计算上可行的先前分布的家属。因此,现有方法采用元学习限制性的对数结构,严重限制其表达性和性能。为绕过这些问题,我们从实用的贝耶斯神经网络推理的角度看待元学习,将以前视为一种随机过程,并在功能空间中进行推断。具体地说,我们将元培训任务视为数据生成过程的样本,将元学习正规化的元学习视为实证性地评估这一随机过程的法律。我们的方法可以通过元学习数据生成过程边缘的分数功能而不是参数空间的先期,来无缝合地获取和代表复杂的先前知识。在一项全面的基准中,我们证明我们的方法在预测准确性和不确定性估计质量方面达到了最先进的业绩。