The potential of graph convolutional neural networks for the task of zero-shot learning has been demonstrated recently. These models are highly sample efficient as related concepts in the graph structure share statistical strength allowing generalization to new classes when faced with a lack of data. However, knowledge from distant nodes can get diluted when propagating through intermediate nodes, because current approaches to zero-shot learning use graph propagation schemes that perform Laplacian smoothing at each layer. We show that extensive smoothing does not help the task of regressing classifier weights in zero-shot learning. In order to still incorporate information from distant nodes and utilize the graph structure, we propose an Attentive Dense Graph Propagation Module (ADGPM). ADGPM allows us to exploit the hierarchical graph structure of the knowledge graph through additional connections. These connections are added based on a node's relationship to its ancestors and descendants and an attention scheme is further used to weigh their contribution depending on the distance to the node. Finally, we illustrate that finetuning of the feature representation after training the ADGPM leads to considerable improvements. Our method achieves competitive results, outperforming previous zero-shot learning approaches.
翻译:这些模型具有高度的样本效率,因为图表结构中的相关概念在缺乏数据的情况下可以共享统计力量,从而能够向新类别推广。然而,在通过中间节点传播时,远程节点的知识可能会被稀释,因为目前对零点学习使用图的传播计划采取的办法,在每一层平滑拉平滑。我们表明,广泛的平滑无助于在零点学习中递减分类权重的任务。为了继续纳入远程节点的信息并利用图表结构,我们提议采用“加速图示促进模块(ADGPM)。ADGPM允许我们通过更多连接来利用知识图的层次图结构。这些联系是根据节点与其祖先和后代的关系而增加的,并且根据节点的距离进一步使用关注计划来权衡其贡献。最后,我们说明在培训了DGPM之后对特征的表述进行了微调,从而取得了相当大的改进。我们的方法实现了竞争性结果,超越了以前的零点学习方法。