Self-supervised learning (SSL) has been extensively explored in recent years. Particularly, generative SSL has seen emerging success in natural language processing and other fields, such as the wide adoption of BERT and GPT. Despite this, contrastive learning--which heavily relies on structural data augmentation and complicated training strategies--has been the dominant approach in graph SSL, while the progress of generative SSL on graphs, especially graph autoencoders (GAEs), has thus far not reached the potential as promised in other fields. In this paper, we identify and examine the issues that negatively impact the development of GAEs, including their reconstruction objective, training robustness, and error metric. We present a masked graph autoencoder GraphMAE that mitigates these issues for generative self-supervised graph learning. Instead of reconstructing structures, we propose to focus on feature reconstruction with both a masking strategy and scaled cosine error that benefit the robust training of GraphMAE. We conduct extensive experiments on 21 public datasets for three different graph learning tasks. The results manifest that GraphMAE--a simple graph autoencoder with our careful designs--can consistently generate outperformance over both contrastive and generative state-of-the-art baselines. This study provides an understanding of graph autoencoders and demonstrates the potential of generative self-supervised learning on graphs.
翻译:近年来,人们广泛探索了自我监督学习(SSL),特别是在自然语言处理和其他领域,如广泛采用BERT和GPT等,自导自导学习(SSL)取得了新成功。尽管如此,大量依赖结构数据增强和复杂培训战略的反向学习(SSL)一直是图SL的主要方法,而图纸上的自导性SSL的进展,特别是图形自导自动编码器(GAE),迄今尚未达到其他领域所承诺的潜力。在本文中,我们发现并研究了对GAE的发展产生不利影响的问题,包括它们的重建目标、培训坚固度和误差度测量。我们展示了一个蒙面图的自动编码图图图图解(Outifical Conformational Conformae),用一个细致的自我监督式图表图解析模型(Outical-Development of the the compatical-Developmental-commogrational-destruction)的模型,我们建议重点重塑结构,而不是重建结构模型。