There is increasing evidence suggesting neural networks' sensitivity to distribution shifts, so that research on out-of-distribution (OOD) generalization comes into the spotlight. Nonetheless, current endeavors mostly focus on Euclidean data, and its formulation for graph-structured data is not clear and remains under-explored, given the two-fold fundamental challenges: 1) the inter-connection among nodes in one graph, which induces non-IID generation of data points even under the same environment, and 2) the structural information in the input graph, which is also informative for prediction. In this paper, we formulate the OOD problem for node-level prediction on graphs and develop a new domain-invariant learning approach, named Explore-to-Extrapolate Risk Minimization, that facilitates GNNs to leverage invariant graph features for prediction. The key difference to existing invariant models is that we design multiple context explorers (specified as graph editers in our case) that are adversarially trained to maximize the variance of risks from multiple virtual environments. Such a design enables the model to extrapolate from a single observed environment which is the common case for node-level prediction. We prove the validity of our method by theoretically showing its guarantee of a valid OOD solution and further demonstrate its power on various real-world datasets for handling distribution shifts from artificial spurious features, cross-domain transfers and dynamic graph evolution.
翻译:越来越多的证据表明神经网络对分布变化的敏感度,因此,关于分配外(OOD)一般化的研究成为焦点。然而,目前的努力主要侧重于欧洲-cliidean数据,而其图表结构数据的设计并不明确,而且仍然未得到充分探讨,因为有两个基本挑战:(1) 同一图中节点之间的相互联系,这导致即使在同一个环境中也非二维生成数据点;(2) 输入图中的结构性信息,它也是用于预测的信息。在本文中,我们为图表上的节偏水平预测设计OOOOD问题,并开发一种新的域-异性学习方法,名为Exploree-to-Extrapolate 风险最小化,这有利于GNNNS利用变异图特性进行预测。与现有变异模型的主要区别在于,我们设计了多种背景勘探者(我们的情况是作为图形编辑者),进行敌对性培训以尽量扩大来自多个虚拟环境的风险差异。在本文中,这种设计使得模型能够从一个观察到的动态变化的模型到从一个观察到的图像处理水平的外推算出其真实的动态变化,这是我们所观察到的动态数据水平上的一个共同的论证。