Learning with limited data is a key challenge for visual recognition. Few-shot learning methods address this challenge by learning an instance embedding function from seen classes and apply the function to instances from unseen classes with limited labels. This style of transfer learning is task-agnostic: the embedding function is not learned optimally discriminative with respect to the unseen classes, where discerning among them is the target task. In this paper, we propose a novel approach to adapt the embedding model to the target classification task, yielding embeddings that are task-specific and are discriminative. To this end, we employ a type of self-attention mechanism called Transformer to transform the embeddings from task-agnostic to task-specific by focusing on relating instances from the test instances to the training instances in both seen and unseen classes. Our approach also extends to both transductive and generalized few-shot classification, two important settings that have essential use cases. We verify the effectiveness of our model on two standard benchmark few-shot classification datasets --- MiniImageNet and CUB, where our approach demonstrates state-of-the-art empirical performance.
翻译:以有限数据进行学习是目视识别的关键挑战。 少见的学习方法通过从可见的班级学习一个实例嵌入功能,并将功能应用到标签有限的无形班级的事例来应对这一挑战。 这种转移学习的风格是任务不可知的: 嵌入功能没有在无形班级中最优的区别性学习, 目标任务就是在这些班级中辨别它们。 在本文件中, 我们提出一种新的方法, 使嵌入模式适应目标分类任务, 产生特定任务和具有歧视性的嵌入。 为此, 我们使用一种自留机制, 叫做变换器, 将嵌入从任务不可知性到任务特定班级的嵌入转换为任务特定。 我们的方法还扩大到了介于测试案例和普通的少发分类, 有两个重要的应用案例。 我们核查了我们两个标准基准的模型的实效, 即微光分分类数据集 -- MinimageNet 和 CUB, 我们的方法展示了当前的经验性表现。