Spectral graph neural networks (GNNs) learn graph representations via spectral-domain graph convolutions. However, most existing spectral graph filters are scalar-to-scalar functions, i.e., mapping a single eigenvalue to a single filtered value, thus ignoring the global pattern of the spectrum. Furthermore, these filters are often constructed based on some fixed-order polynomials, which have limited expressiveness and flexibility. To tackle these issues, we introduce Specformer, which effectively encodes the set of all eigenvalues and performs self-attention in the spectral domain, leading to a learnable set-to-set spectral filter. We also design a decoder with learnable bases to enable non-local graph convolution. Importantly, Specformer is equivariant to permutation. By stacking multiple Specformer layers, one can build a powerful spectral GNN. On synthetic datasets, we show that our Specformer can better recover ground-truth spectral filters than other spectral GNNs. Extensive experiments of both node-level and graph-level tasks on real-world graph datasets show that our Specformer outperforms state-of-the-art GNNs and learns meaningful spectrum patterns. Code and data are available at https://github.com/bdy9527/Specformer.
翻译:光谱图像神经网络(GNNSs)通过光谱-多角度图像图解图解,通过光谱-多角度图谱图象变相,学习图解图解。然而,大多数现有的光谱图像过滤器是星际到星际的功能,即将单一的亚元值映射成单一过滤值,从而忽略光谱的全局模式。此外,这些过滤器通常建筑在固定序列多元多面体的基础上,这些多面体的表达性和灵活性有限。为了解决这些问题,我们引入了Specdyex,它有效地编码了所有电子值的集,并在光谱域中进行自我注意,导致一个可学习的集成光谱过滤器。我们还设计了一个带有可学习模式基础的解码器,以使非局部图象变相法成为全局性。通过堆积多谱层,可以建立强大的光谱GNNNN。在合成数据设置上,我们Spetrafors能够比其他的光谱-level谱-travel-travel-lex-legs GNNNNS-cods-colm的GNPNS-CS-CS-C-C-C-C-C-C-C-CRD-C-D-CRD-S-C-S-S-C-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-D-D-SDFD-S-S-SD-SD-S-S-SD-S-D-S-S-S-S-S-S-SD-S-S-S-S-S-S-S-S-S-S-S-D-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-D-S-S-D-D-</s>