Few-Shot Learning (FSL) is a challenging task, i.e., how to recognize novel classes with few examples? Pre-training based methods effectively tackle the problem by pre-training a feature extractor and then predict novel classes via a nearest neighbor classifier with mean-based prototypes. Nevertheless, due to the data scarcity, the mean-based prototypes are usually biased. In this paper, we diminish the bias by regarding it as a prototype optimization problem. Although the existing meta-optimizers can also be applied for the optimization, they all overlook a crucial gradient bias issue, i.e., the mean-based gradient estimation is also biased on scarce data. Consequently, we regard the gradient itself as meta-knowledge and then propose a novel prototype optimization-based meta-learning framework, called MetaNODE. Specifically, we first regard the mean-based prototypes as initial prototypes, and then model the process of prototype optimization as continuous-time dynamics specified by a Neural Ordinary Differential Equation (Neural ODE). A gradient flow inference network is carefully designed to learn to estimate the continuous gradients for prototype dynamics. Finally, the optimal prototypes can be obtained by solving the Neural ODE using the Runge-Kutta method. Extensive experiments demonstrate that our proposed method obtains superior performance over the previous state-of-the-art methods. Our code will be publicly available upon acceptance.
翻译:低热学习( FSL) 是一项艰巨的任务, 即如何以几个例子来识别新类? 培训前方法通过先训练一个地物提取器来有效解决问题, 然后通过一个离近的邻居分类器, 以平均原型来预测新类。 然而, 由于数据稀缺, 以平均值为基础的原型通常有偏差。 在本文中, 我们通过将原型视为原型优化问题来减少偏见。 尽管现有的元优化器也可以应用到优化中, 但它们都忽略了关键的梯度偏差问题, 即平均梯度估计也偏向于稀缺的数据。 因此, 我们把梯度本身视为元知识, 然后提出一个新的基于优化的原型元学习框架, 称为MetaNODE。 具体地说, 我们首先将原型原型原型原型视为初始原型原型, 然后将原型优化过程作为由神经质普通差异化( Neuroral Squal Equal Equalation) 指定的连续时间动态。 梯流网络经过仔细设计, 以便学习如何利用我们提出的前期的高级实验方法, 展示我们的最佳原型模型, 将获得最佳原型模型。