Despite its outstanding performance in various graph tasks, vanilla Message Passing Neural Network (MPNN) usually fails in link prediction tasks, as it only uses representations of two individual target nodes and ignores the pairwise relation between them. To capture the pairwise relations, some models add manual features to the input graph and use the output of MPNN to produce pairwise representations. In contrast, others directly use manual features as pairwise representations. Though this simplification avoids applying a GNN to each link individually and thus improves scalability, these models still have much room for performance improvement due to the hand-crafted and unlearnable pairwise features. To upgrade performance while maintaining scalability, we propose Neural Common Neighbor (NCN), which uses learnable pairwise representations. To further boost NCN, we study the unobserved link problem. The incompleteness of the graph is ubiquitous and leads to distribution shifts between the training and test set, loss of common neighbor information, and performance degradation of models. Therefore, we propose two intervention methods: common neighbor completion and target link removal. Combining the two methods with NCN, we propose Neural Common Neighbor with Completion (NCNC). NCN and NCNC outperform recent strong baselines by large margins. NCNC achieves state-of-the-art performance in link prediction tasks.
翻译:尽管在各种图表任务中表现出色,但香草信息传递神经网络(MPNN)通常在预测任务上没有成功,因为它只使用两个单个目标节点的表示方式,而忽略了它们之间的对称关系。为了捕捉对称关系,一些模型在输入图中添加人工特征,并利用MPNN的输出产生对称代表方式。相比之下,另一些模型则直接使用手工特征作为对称表达方式。虽然这种简化避免对每个链接都应用GNN,从而改进了可缩放性,但这些模型由于手制和不可忽略的对称性功能,仍然有很大的改进空间。为了提高性能,我们在保持可缩放性的同时,我们提议采用Nural Common Common Common Neghbbor(NCN),它使用可学习的对称代表方式。为了进一步提升NCNNCN,我们研究未观察到的链接问题。图表的不完全性,导致培训和测试设置之间的分布变化,失去共同的邻域信息,以及模型的性能退化。因此,我们建议两种干预方法:共同的邻居完成和目标链接,同时将NCNNCNCNCM的两种方法与NCM的大规模基线连接。我们提议通过NCMNC的大规模业绩基准,我们建议将NCFMMMMM的大规模升级的进度连接。