Generalizing a pretrained model to unseen datasets without retraining is an essential step toward a foundation model. However, achieving such cross-dataset, fully inductive inference is difficult in graph-structured data where feature spaces vary widely in both dimensionality and semantics. Any transformation in the feature space can easily violate the inductive applicability to unseen datasets, strictly limiting the design space of a graph model. In this work, we introduce the view space, a novel representational axis in which arbitrary graphs can be naturally encoded in a unified manner. We then propose Graph View Transformation (GVT), a node- and feature-permutation-equivariant mapping in the view space. GVT serves as the building block for Recurrent GVT, a fully inductive model for node representation learning. Pretrained on OGBN-Arxiv and evaluated on 27 node-classification benchmarks, Recurrent GVT outperforms GraphAny, the prior fully inductive graph model, by +8.93% and surpasses 12 individually tuned GNNs by at least +3.30%. These results establish the view space as a principled and effective ground for fully inductive node representation learning.
翻译:将预训练模型推广到未见过的数据集而无需重新训练,是实现基础模型的关键步骤。然而,在图结构数据中实现这种跨数据集的完全归纳推理十分困难,因为特征空间在维度和语义上均存在显著差异。特征空间的任何变换都可能轻易破坏对未见数据集的归纳适用性,从而严格限制了图模型的设计空间。本文引入视图空间这一新颖的表示轴,其中任意图均可自然地以统一方式编码。随后,我们提出图视图变换(GVT),一种在视图空间中具有节点与特征置换等变性的映射。GVT作为循环GVT的基础构建模块,后者是一种用于节点表示学习的完全归纳模型。通过在OGBN-Arxiv上预训练并在27个节点分类基准上评估,循环GVT以+8.93%的优势超越了先前完全归纳图模型GraphAny,并至少以+3.30%的优势优于12个单独调优的图神经网络。这些结果表明,视图空间为完全归纳节点表示学习提供了原理性且有效的理论基础。