Transformers have achieved remarkable performance in a myriad of fields including natural language processing and computer vision. However, when it comes to the graph mining area, where graph neural network (GNN) has been the dominant paradigm, transformers haven't achieved competitive performance, especially on the node classification task. Existing graph transformer models typically adopt fully-connected attention mechanism on the whole input graph and thus suffer from severe scalability issues and are intractable to train in data insufficient cases. To alleviate these issues, we propose a novel Gophormer model which applies transformers on ego-graphs instead of full-graphs. Specifically, Node2Seq module is proposed to sample ego-graphs as the input of transformers, which alleviates the challenge of scalability and serves as an effective data augmentation technique to boost model performance. Moreover, different from the feature-based attention strategy in vanilla transformers, we propose a proximity-enhanced attention mechanism to capture the fine-grained structural bias. In order to handle the uncertainty introduced by the ego-graph sampling, we further propose a consistency regularization and a multi-sample inference strategy for stabilized training and testing, respectively. Extensive experiments on six benchmark datasets are conducted to demonstrate the superiority of Gophormer over existing graph transformers and popular GNNs, revealing the promising future of graph transformers.
翻译:包括自然语言处理和计算机视觉在内的众多领域,变异器都取得了显著的绩效。然而,在图形采矿领域,图形神经网络(GNNN)一直是主导模式,但变异器没有取得竞争力,特别是在节点分类任务方面。现有的图形变异器模型通常在整个投入图中采用完全连接的注意机制,因此受到严重可缩缩缩缩问题的影响,因此难以在数据不足的情况下进行数据培训。为了缓解这些问题,我们提议了一个新型的Gophormer模型,将变异器应用在自我图上而不是全文上。具体地说,将Node2Seq模块推荐给自我图样本,作为变异器的投入,这缓解了可缩缩缩的挑战,并成为提高模型性的有效数据增强技术。此外,与香草变变变变器基于特征的注意战略不同,我们提议了一种近似强化的注意机制,以捕捉精细的结构性偏差。为了处理自我图取样带来的不确定性,我们进一步提议将Nde2Seq模块作为自我图的样本,作为变异的样本,作为变异器的样本。Gophregalimerveriforglegregal 测试分别进行。Ggoprismaregregregregregregregregregregregregal的Gs