Text-attributed graphs require models to effectively combine strong textual understanding with structurally informed reasoning. Existing approaches either rely on GNNs--limited by over-smoothing and hop-dependent diffusion--or employ Transformers that overlook graph topology and treat nodes as isolated sequences. We propose Odin (Oriented Dual-module INtegration), a new architecture that injects graph structure into Transformers at selected depths through an oriented dual-module mechanism.Unlike message-passing GNNs, Odin does not rely on multi-hop diffusion; instead, multi-hop structures are integrated at specific Transformer layers, yielding low-, mid-, and high-level structural abstraction aligned with the model's semantic hierarchy. Because aggregation operates on the global [CLS] representation, Odin fundamentally avoids over-smoothing and decouples structural abstraction from neighborhood size or graph topology. We further establish that Odin's expressive power strictly contains that of both pure Transformers and GNNs.To make the design efficient in large-scale or low-resource settings, we introduce Light Odin, a lightweight variant that preserves the same layer-aligned structural abstraction for faster training and inference. Experiments on multiple text-rich graph benchmarks show that Odin achieves state-of-the-art accuracy, while Light Odin delivers competitive performance with significantly reduced computational cost. Together, Odin and Light Odin form a unified, hop-free framework for principled structure-text integration. The source code of this model has been released at https://github.com/hongkaifeng/Odin.
翻译:文本属性图要求模型能够有效结合强大的文本理解能力与结构感知推理。现有方法要么依赖图神经网络(GNNs)——受限于过度平滑和跳数依赖的扩散机制,要么采用忽略图拓扑结构、将节点视为孤立序列的Transformer模型。本文提出Odin(定向双模块集成),一种通过定向双模块机制在特定深度将图结构注入Transformer的新架构。与基于消息传递的GNNs不同,Odin不依赖多跳扩散;相反,多跳结构被集成到特定的Transformer层中,生成与模型语义层次对齐的低、中、高级结构抽象。由于聚合操作作用于全局[CLS]表示,Odin从根本上避免了过度平滑问题,并将结构抽象与邻域大小或图拓扑解耦。我们进一步证明Odin的表达能力严格包含纯Transformer和GNNs的表达能力。为在大规模或低资源场景下实现高效设计,我们引入了Light Odin,这是一种轻量级变体,保留了相同的层对齐结构抽象,以实现更快的训练和推理。在多个文本丰富图基准测试上的实验表明,Odin实现了最先进的准确率,而Light Odin以显著降低的计算成本提供了具有竞争力的性能。Odin与Light Odin共同构成了一个统一的、无需跳数依赖的结构-文本集成理论框架。该模型的源代码已发布于https://github.com/hongkaifeng/Odin。