In this work, we propose BertGCN, a model that combines large scale pretraining and transductive learning for text classification. BertGCN constructs a heterogeneous graph over the dataset and represents documents as nodes using BERT representations. By jointly training the BERT and GCN modules within BertGCN, the proposed model is able to leverage the advantages of both worlds: large-scale pretraining which takes the advantage of the massive amount of raw data and transductive learning which jointly learns representations for both training data and unlabeled test data by propagating label influence through graph convolution. Experiments show that BertGCN achieves SOTA performances on a wide range of text classification datasets. Code is available at https://github.com/ZeroRin/BertGCN.
翻译:在这项工作中,我们提议BertGCN,这是一个将大规模预培训和转导学习结合起来用于文本分类的模型,BertGCN在数据集上制作了一个多式图表,并使用BERT代表作为节点代表文件,通过在BertGCN内联合培训BERT模块和GCN模块,拟议的模型能够利用两个世界的优势:利用大量原始数据和转导学习的大规模预培训,通过通过图解组合传播标签影响,共同学习培训数据和未贴标签测试数据的表现。实验显示,BertGCN在广泛的文本分类数据集中取得了SOTA性能,可在https://github.com/ZeroRin/BertGCN上查阅代码。