Graph Neural Networks (GNNs) have been widely used on graph data and have shown exceptional performance in the task of link prediction. Despite their effectiveness, GNNs often suffer from high latency due to non-trivial neighborhood data dependency in practical deployments. To address this issue, researchers have proposed methods based on knowledge distillation (KD) to transfer the knowledge from teacher GNNs to student MLPs, which are known to be efficient even with industrial scale data, and have shown promising results on node classification. Nonetheless, using KD to accelerate link prediction is still unexplored. In this work, we start with exploring two direct analogs of traditional KD for link prediction, i.e., predicted logit-based matching and node representation-based matching. Upon observing direct KD analogs do not perform well for link prediction, we propose a relational KD framework, Linkless Link Prediction (LLP). Unlike simple KD methods that match independent link logits or node representations, LLP distills relational knowledge that is centered around each (anchor) node to the student MLP. Specifically, we propose two matching strategies that complement each other: rank-based matching and distribution-based matching. Extensive experiments demonstrate that LLP boosts the link prediction performance of MLPs with significant margins, and even outperforms the teacher GNNs on 6 out of 9 benchmarks. LLP also achieves a 776.37x speedup in link prediction inference compared to GNNs on the large scale OGB-Citation2 dataset.
翻译:图表神经网络(GNNs)在图表数据中被广泛使用,显示在连接预测任务中表现得非常出色。尽管这些网络具有效力,但GNNs往往由于在实际部署中非三角邻里数据依赖性而处于高度悬浮状态。为解决这一问题,研究人员提出了基于知识蒸馏(KD)的方法,将知识从教师GNNs(KD)转移到学生MLP(即使与工业规模数据相比也已知效率很高,在节点分类上也显示出有希望的结果。尽管如此,在这项工作中,使用KD(加速连接预测)仍然没有得到探索。我们开始探索传统KD的两个直接的直线链接,用于链接预测,即预测基于日志的匹配和基于节点的表示匹配匹配匹配。在直接观测KD类模拟时,我们提议一个关系KD框架,即即使与工业规模数据相匹配,也与基于7P值的单位或节点表示相匹配的简单KD方法不同,LP在每一个(anchor)左右的师级点上都与高级双级LP(OLP) 匹配。我们提议在一次大级预测中要比高级GLP(S-LP) 。