Meta-learning extracts the common knowledge acquired from learning different tasks and uses it for unseen tasks. It demonstrates a clear advantage on tasks that have insufficient training data, e.g., few-shot learning. In most meta-learning methods, tasks are implicitly related via the shared model or optimizer. In this paper, we show that a meta-learner that explicitly relates tasks on a graph describing the relations of their output dimensions (e.g., classes) can significantly improve the performance of few-shot learning. This type of graph is usually free or cheap to obtain but has rarely been explored in previous works. We study the prototype based few-shot classification, in which a prototype is generated for each class, such that the nearest neighbor search between the prototypes produces an accurate classification. We introduce "Gated Propagation Network (GPN)", which learns to propagate messages between prototypes of different classes on the graph, so that learning the prototype of each class benefits from the data of other related classes. In GPN, an attention mechanism is used for the aggregation of messages from neighboring classes, and a gate is deployed to choose between the aggregated messages and the message from the class itself. GPN is trained on a sequence of tasks from many-shot to few-shot generated by subgraph sampling. During training, it is able to reuse and update previously achieved prototypes from the memory in a life-long learning cycle. In experiments, we change the training-test discrepancy and test task generation settings for thorough evaluations. GPN outperforms recent meta-learning methods on two benchmark datasets in all studied cases.
翻译:元化学习提取了从学习不同任务中获得的共同知识,并将其用于无形任务。 它展示了在培训数据不足的任务上的一个明显优势, 例如, 少见的学习。 在大多数元化学习方法中, 任务通过共享模型或优化器暗含关联。 在本文中, 我们显示, 将任务明确连接到描述其输出维度关系( 例如, 类) 的图表上, 可以大大改进微小学习的绩效。 这种图表通常免费或廉价, 但却在以往的工作中很少探索。 我们研究了基于原型的少见分类, 其中为每类生成了一个原型, 从而产生了一个原型的元化的分类。 在多数元的元化学习方法中, 使用基于原型的少见的分类, 也就是基于原型的原型的原型( 原型 ), 使最近的邻居搜索方法产生精确的分类。 我们引入了“ GGPGN ”, 在图表上, 学习不同类的原型, 学习每个类的原型的原型, 使用每个类的原型的原型的原型, 注意机制用来收集信息,, 和从最近版本的校订的试制的样本中, 学习中, 从几类的实验的实验的实验的 更新。