Graph Neural Networks (GNNs) are de facto solutions to structural data learning. However, it is susceptible to low-quality and unreliable structure, which has been a norm rather than an exception in real-world graphs. Existing graph structure learning (GSL) frameworks still lack robustness and interpretability. This paper proposes a general GSL framework, SE-GSL, through structural entropy and the graph hierarchy abstracted in the encoding tree. Particularly, we exploit the one-dimensional structural entropy to maximize embedded information content when auxiliary neighbourhood attributes are fused to enhance the original graph. A new scheme of constructing optimal encoding trees is proposed to minimize the uncertainty and noises in the graph whilst assuring proper community partition in hierarchical abstraction. We present a novel sample-based mechanism for restoring the graph structure via node structural entropy distribution. It increases the connectivity among nodes with larger uncertainty in lower-level communities. SE-GSL is compatible with various GNN models and enhances the robustness towards noisy and heterophily structures. Extensive experiments show significant improvements in the effectiveness and robustness of structure learning and node representation learning.
翻译:图神经网络(GNN)是结构数据学习的事实解决方案。然而,存在低质量和不可靠结构,这在实际图中已经成为惯例而非例外。现有的图结构学习(GSL)框架仍然缺乏鲁棒性和可解释性。本文提出了一个通用的GSL框架SE-GSL,通过结构熵和编码树中抽象的图层次,特别是利用一维结构熵,最大化融合辅助邻域属性以增强原图时嵌入信息内容。提出了一种构建最优编码树的新方案,以在保证适当的层次抽象中最小化图中的不确定性和噪声并确保社区划分。我们提出了一种新颖的基于样本的机制,通过节点结构熵分布还原图结构。它增加了底层社区中不确定度较大的节点之间的连通性。SE-GSL与各种GNN模型兼容,增强了对嘈杂和异质性结构的鲁棒性。广泛的实验显示了结构学习和节点表示学习的有效性和鲁棒性的显着提高。