Graph neural networks (GNN) have recently emerged as a vehicle for applying deep network architectures to graph and relational data. However, given the increasing size of industrial datasets, in many practical situations the message passing computations required for sharing information across GNN layers are no longer scalable. Although various sampling methods have been introduced to approximate full-graph training within a tractable budget, there remain unresolved complications such as high variances and limited theoretical guarantees. To address these issues, we build upon existing work and treat GNN neighbor sampling as a multi-armed bandit problem but with a newly-designed reward function that introduces some degree of bias designed to reduce variance and avoid unstable, possibly-unbounded pay outs. And unlike prior bandit-GNN use cases, the resulting policy leads to near-optimal regret while accounting for the GNN training dynamics introduced by SGD. From a practical standpoint, this translates into lower variance estimates and competitive or superior test accuracy across several benchmarks.
翻译:最近,由于工业数据集规模不断扩大,在很多实际情况下,在GNN层之间共享信息所需的信息传递计算方法不再可以伸缩。虽然采用了各种抽样方法,在可移动预算范围内大致进行全图培训,但仍有一些未解决的复杂问题,如差异大和理论保障有限。为了解决这些问题,我们以现有工作为基础,将GNN邻居取样视为多臂土匪问题,但新设计的奖励功能带有某种程度的偏差,目的是减少差异,避免不稳定的、可能不受约束的薪酬外溢。与以往的bandit-GNN使用案例不同,由此产生的政策导致近乎最佳的遗憾,同时核算SGD引入的GNN培训动态。从实际角度讲,这转化为差异估计较低,在几个基准上具有竞争性或更高的测试精度。