Few-shot learning is a challenging problem that requires a model to recognize novel classes with few labeled data. In this paper, we aim to find the expected prototypes of the novel classes, which have the maximum cosine similarity with the samples of the same class. Firstly, we propose a cosine similarity based prototypical network to compute basic prototypes of the novel classes from the few samples. A bias diminishing module is further proposed for prototype rectification since the basic prototypes computed in the low-data regime are biased against the expected prototypes. In our method, the intra-class bias and the cross-class bias are diminished to modify the prototypes. Then we give a theoretical analysis of the impact of the bias diminishing module on the expected performance of our method. We conduct extensive experiments on four few-shot benchmarks and further analyze the advantage of the bias diminishing module. The bias diminishing module brings in significant improvement by a large margin of 3% to 9% in general. Notably, our approach achieves state-of-the-art performance on miniImageNet (70.31% in 1-shot and 81.89% in 5-shot) and tieredImageNet (78.74% in 1-shot and 86.92% in 5-shot), which demonstrates the superiority of the proposed method.
翻译:少见的学习是一个具有挑战性的问题,需要模型来识别带有少量标签数据的新类。 在本文中,我们的目标是找到新类的预期原型,这些原型与同类样本具有最大相似性。 首先,我们提议建立一个基于直径相似的原型网络,以从少数样本中计算新类的基本原型。 进一步提议为原型的校正提出一个缩小偏差模块,因为在低数据制度中计算的基本原型与预期原型有偏差。 在我们的方法中,本级内部偏差和跨级偏差会减少,以修改原型。 然后我们从理论上分析缩小偏差模块对本方法预期性能的影响。 我们对四张微小基准进行广泛的实验,并进一步分析缩小新类原型原型的优点。 缩小偏差模块使原型大大改进,总幅度为3%至9%。 值得注意的是,我们的方法在微型ImageNet上取得了最先进的性能(0.31和81.89%的原型偏差性偏差,5张图中为5-74%)和一级INet方法展示了1-74%的领先率。