Learning on graphs, where instance nodes are inter-connected, has become one of the central problems for deep learning, as relational structures are pervasive and induce data inter-dependence which hinders trivial adaptation of existing approaches that assume inputs to be i.i.d.~sampled. However, current models mostly focus on improving testing performance of in-distribution data and largely ignore the potential risk w.r.t. out-of-distribution (OOD) testing samples that may cause negative outcome if the prediction is overconfident on them. In this paper, we investigate the under-explored problem, OOD detection on graph-structured data, and identify a provably effective OOD discriminator based on an energy function directly extracted from graph neural networks trained with standard classification loss. This paves a way for a simple, powerful and efficient OOD detection model for GNN-based learning on graphs, which we call GNNSafe. It also has nice theoretical properties that guarantee an overall distinguishable margin between the detection scores for in-distribution and OOD samples, which, more critically, can be further strengthened by a learning-free energy belief propagation scheme. For comprehensive evaluation, we introduce new benchmark settings that evaluate the model for detecting OOD data from both synthetic and real distribution shifts (cross-domain graph shifts and temporal graph shifts). The results show that GNNSafe achieves up to $17.0\%$ AUROC improvement over state-of-the-arts and it could serve as simple yet strong baselines in such an under-developed area.
翻译:在图表上学习是互为关联的,因此已经成为深层学习的中心问题之一,因为关系结构十分普遍,并导致数据相互依存,从而阻碍对现有方法的微小调整,这些方法假定投入是按标准分类损失直接从经过培训的图形神经网络中提取的能源功能,但是,目前的模型主要侧重于改进分布中数据测试性能,并基本上忽视在分布中测试可能带来负面结果的潜在风险(OOOD),我们称之为GNNSafe。在本文中,我们调查探索不足的问题,OOOD对图表结构数据进行OOOD检测,并找出一种基于直接从经过标准分类损失培训的图形神经网络中提取的能函数的、可被证实有效的OOD歧视者。这为GNNN在图表上学习一个简单、有力和有效的 OOOOD检测模式,我们称之为GNNSafe。它还具有良好的理论性能,可以保证在分配中和OODD基准值的检测分数之间有一个总体可分辨的差差值。 更为关键的是,我们可以更精确地通过学习的OOD全面的变化来进一步评估新的OD数据结构,从而显示对GROD的模型进行真正的变化。