Discriminative pre-trained language models (PLMs) learn to predict original texts from intentionally corrupted ones. Taking the former text as positive and the latter as negative samples, the PLM can be trained effectively for contextualized representation. However, the training of such a type of PLMs highly relies on the quality of the automatically constructed samples. Existing PLMs simply treat all corrupted texts as equal negative without any examination, which actually lets the resulting model inevitably suffer from the false negative issue where training is carried out on pseudo-negative data and leads to less efficiency and less robustness in the resulting PLMs. In this work, on the basis of defining the false negative issue in discriminative PLMs that has been ignored for a long time, we design enhanced pre-training methods to counteract false negative predictions and encourage pre-training language models on true negatives by correcting the harmful gradient updates subject to false negative predictions. Experimental results on GLUE and SQuAD benchmarks show that our counter-false-negative pre-training methods indeed bring about better performance together with stronger robustness.
翻译:语言模型(PLM)学会从有意腐蚀的语言模型中预测原始文本。用前一文本为正数,而后一文本为负数样本,PLM可以有效地进行背景代表培训。然而,对此类类型的PLM的培训高度依赖自动制成样本的质量。现有的PLM只是将所有腐败文本视为等同的负数,而不经过任何检查,这实际上使由此产生的模式不可避免地受到虚假的负面问题的影响,因为就伪阴性数据进行培训,从而降低培训前PLM的效率和强度。在这项工作中,根据长期以来被忽视的歧视性的PLM的虚假负面问题的定义,我们设计了强化的培训前方法,以抵制虚假的负面预测,并鼓励对真实负面的预先语言模型进行培训,纠正有害的梯度更新,但需受到虚假的负面预测。GLUE和SQUAD基准的实验结果表明,我们的反反反阴性培训前方法确实能更好,同时加强力度。