As one of the most popular GNN architectures, the graph attention networks (GAT) is considered the most advanced learning architecture for graph representation and has been widely used in various graph mining tasks with impressive results. However, since GAT was proposed, none of the existing studies have provided systematic insight into the relationship between the performance of GAT and the number of layers, which is a critical issue in guiding model performance improvement. In this paper, we perform a systematic experimental evaluation and based on the experimental results, we find two important facts: (1) the main factor limiting the accuracy of the GAT model as the number of layers increases is the oversquashing phenomenon; (2) among the previous improvements applied to the GNN model, only the residual connection can significantly improve the GAT model performance. We combine these two important findings to provide a theoretical explanation that it is the residual connection that mitigates the loss of original feature information due to oversquashing and thus improves the deep GAT model performance. This provides empirical insights and guidelines for researchers to design the GAT variant model with appropriate depth and well performance. To demonstrate the effectiveness of our proposed guidelines, we propose a GAT variant model-ADGAT that adaptively selects the number of layers based on the sparsity of the graph, and experimentally demonstrate that the effectiveness of our model is significantly improved over the original GAT.
翻译:作为最受欢迎的GNN结构之一,石图关注网络被认为是用于图形代表的最先进的学习架构,在各种图形采矿任务中被广泛使用,并取得了令人印象深刻的成果;然而,自提议采用GAT以来,现有研究没有一项对GAT绩效和层数之间的关系提供系统深入的了解,这是指导模型绩效改进的一个关键问题。在本文件中,我们进行了系统的实验性评估,并根据实验结果,发现两个重要事实:(1) 层数增加是过度夸大现象,因此限制GAT模型准确性的主要因素;(2) 在以前对GNN模型采用的改进中,只有剩余连接才能大大改善GAT模型的绩效。我们把这些重要研究结果结合起来,从理论上解释,这是减轻因夸大和从而改进GAT模型深度和良好业绩而导致原始特征信息损失的剩余联系。这为研究人员设计GAT变异模型提供了经验性模型和准则,以适当深度和良好性能的方式设计GAT模型模型。为了证明我们提议的指南的有效性,我们提议在GAT模型基础上大幅改进GAT变异性模型和GAT模型,我们提议在模型上选择基于原型模型的GAAT模型的模型的模型的模型,以显示我们的适应性。