Learning latent representations of nodes in graphs is an important and ubiquitous task with widespread applications such as link prediction, node classification, and graph visualization. Previous methods on graph representation learning mainly focus on static graphs, however, many real-world graphs are dynamic and evolve over time. In this paper, we present Dynamic Self-Attention Network (DySAT), a novel neural architecture that operates on dynamic graphs and learns node representations that capture both structural properties and temporal evolutionary patterns. Specifically, DySAT computes node representations by jointly employing self-attention layers along two dimensions: structural neighborhood and temporal dynamics. We conduct link prediction experiments on two classes of graphs: communication networks and bipartite rating networks. Our experimental results show that DySAT has a significant performance gain over several different state-of-the-art graph embedding baselines.
翻译:图表中节点的学习潜在表达方式是一项重要且无处不在的任务,具有广泛的应用,如链接预测、节点分类和图形可视化等。以往的图形表达方式学习方法主要侧重于静态图形,然而,许多真实世界的图形是动态的,并随着时间的推移而演变。在本文中,我们展示了动态自控网络(DySAT),这是一个新型神经结构,它以动态图形运作,并学习捕捉结构属性和时间进化模式的节点表达方式。具体地说,DySAT通过在两个维度上联合使用自留层来计算节点表达方式:结构邻里和时间动态。我们对两类图表进行连接预测实验:通信网络和双端评级网络。我们的实验结果表明,DySAT在几个不同的图表嵌入基线上拥有显著的性能收益。