Graph convolutional neural networks have recently shown great potential for the task of zero-shot learning. These models are highly sample efficient as related concepts in the graph structure share statistical strength allowing generalization to new classes when faced with a lack of data. However, multi-layer architectures, which are required to propagate knowledge to distant nodes in the graph, dilute the knowledge by performing extensive Laplacian smoothing at each layer and thereby consequently decrease performance. In order to still enjoy the benefit brought by the graph structure while preventing dilution of knowledge from distant nodes, we propose a Dense Graph Propagation (DGP) module with carefully designed direct links among distant nodes. DGP allows us to exploit the hierarchical graph structure of the knowledge graph through additional connections. These connections are added based on a node's relationship to its ancestors and descendants. A weighting scheme is further used to weigh their contribution depending on the distance to the node to improve information propagation in the graph. Combined with finetuning of the representations in a two-stage training approach our method outperforms state-of-the-art zero-shot learning approaches.
翻译:这些模型具有高度的样本效率,因为图表结构中的相关概念在缺乏数据的情况下可以共享统计力量,从而可以向新类别进行概括化。然而,多层结构需要将知识传播到图中的远节点,而多层结构则需要将知识传播到图中的远端节点,通过在每一层进行广泛的拉普拉西亚平滑,从而降低性能,从而冲淡知识。为了继续享受图结构带来的惠益,同时防止从远端节点淡化知识,我们提议建立一个Dense图表促进模块,在远端节点之间精心设计直接链接。DGP允许我们通过更多连接利用知识图的层次图结构。这些连接是根据节点与其祖先和后代的关系而增加的。还进一步使用加权计划来权衡其贡献,取决于与图中节点的距离,以改进图中的信息传播。同时在两阶段培训方法中,对演示的介绍进行微调,我们的方法优于最先进的零点学习方法。