Knowledge graphs (KGs) facilitate a wide variety of applications due to their ability to store relational knowledge applicable to many areas. Despite great efforts invested in creation and maintenance, even the largest KGs are far from complete. Hence, KG completion (KGC) has become one of the most crucial tasks for KG research. Recently, considerable literature in this space has centered around the use of Graph Neural Networks (GNNs) to learn powerful embeddings which leverage topological structures in the KGs. Specifically, dedicated efforts have been made to extend GNNs, which are commonly designed for simple homogeneous and uni-relational graphs, to the KG context which has diverse and multi-relational connections between entities, by designing more complex aggregation schemes over neighboring nodes (crucial to GNN performance) to appropriately leverage multi-relational information. The success of these methods is naturally attributed to the use of GNNs over simpler multi-layer perceptron (MLP) models, owing to their additional aggregation functionality. In this work, we find that surprisingly, simple MLP models are able to achieve comparable performance to GNNs, suggesting that aggregation may not be as crucial as previously believed. With further exploration, we show careful scoring function and loss function design has a much stronger influence on KGC model performance, and aggregation is not practically required. This suggests a conflation of scoring function design, loss function design, and aggregation in prior work, with promising insights regarding the scalability of state-of-the-art KGC methods today, as well as careful attention to more suitable aggregation designs for KGC tasks tomorrow. The implementation is available online: https://github.com/Juanhui28/Are_MPNNs_helpful.
翻译:知识图形( KGs) 有助于广泛的应用。 尽管在创建和维护方面投入了巨大的努力, 但最大的KGs远未完全完成。 因此, KG的完成( KGC) 已经成为KG研究中最关键的任务之一。 最近, 这个空间的大量文献围绕图形神经网络( GNNS) 学习强大的嵌入, 从而利用KGs28 的表层结构。 具体地说, 已经作出了专门的努力, 将GNNs推广到KGs, 这些GNS通常设计为简单的同质和单一关系图, 甚至最大的KGGGs远没有完成。 因此, KGGGs的完成( GNS 性能的标记) 成为了对多关系网络( GNS) 的恰当利用。 这些方法的成功自然归功于GNS的利用, 将GNS 用于更简单的模型/ MLP( MLP) 的模型, 以及更有前途的汇总功能。 我们发现, 简单的 MLP 模型不是用来用来简单、 多层次的, KNGsurveal 的功能可以比 设计的操作更精确地显示我们更精确的运行。