In this work, we represent CMV-BERT, which improves the pretraining of a language model via two ingredients: (a) contrastive learning, which is well studied in the area of computer vision; (b) multiple vocabularies, one of which is fine-grained and the other is coarse-grained. The two methods both provide different views of an original sentence, and both are shown to be beneficial. Downstream tasks demonstrate our proposed CMV-BERT are effective in improving the pretrained language models.
翻译:在这项工作中,我们代表CMV-BERT,它通过两个要素改进了语言模式的预培训:(a) 对比学习,在计算机视觉领域对此进行了很好的研究;(b) 多种词汇,其中一种是精细的,另一种是粗略的,两种方法都对原句提出了不同的看法,而且两者都证明是有益的。 下游任务表明,我们提议的CMV-BERT在改进预先培训的语言模式方面是有效的。