Pre-trained models have been used in many fields in recent years, ranging from natural language understanding to computer vision and natural language generation. Nowadays, the performance of these natural language generation models is overly dependent on the model's scale and the dataset's size. While the larger language model is excellent in some respects, it cannot learn up-to-date knowledge and is relatively difficult to relearn. In this paper, a new adversarial process learning method is called Auto-Learning, which can improve the performance of any natural language generation model without the help of additional datasets. Auto-Learning includes two models: $G$ is a text generation model, and $D$ can test whether the data generated by G is legitimate. Firstly, the fine-tuned $D$ model is used as the brain's knowledge base before the process. Then the text generated by the $G$ model is used as the input of $D$ to determine whether the text is legitimate. Finally, $G$ is fine-tuned according to the output of $D$. This adversarial process is like a self-escalation of the brain through some a priori knowledge. When this adversarial system wants to learn something new, simply fine-tune the $D$ model. Our approach applies to Autoregressive Language Modeling for all Transformer classes. Auto-Learning enables 8 models to achieve stable improvement in 10 natural language processing tasks without any change in structure.
翻译:近年来,许多领域都采用了预先培训的模型,从自然语言理解到计算机视觉和自然语言生成等,从自然语言理解到计算机视觉和自然语言生成等。如今,这些自然语言生成模型的性能过于依赖模型的规模和数据集的大小。虽然较大的语言模型在某些方面是优秀的,但它无法学习最新知识,而且相对难以再读取。在本文中,一个新的对抗性进程学习方法称为“自动学习”,它可以在没有额外数据集帮助的情况下改进任何自然语言生成模型的性能。自动学习包括两个模型:$G$是一个文本生成模型,而$D美元可以测试G生成的数据是否合法。首先,微调的$D美元模型在程序之前被用作大脑的知识基础。然后,用$G模式生成的文本作为美元的投入,以确定文本是否合法。最后,$G美元可以根据$D的输出进行微调。这一对抗性学习过程就像一个自缩的脑变化模型,通过某种前期的LA级知识来进行。当该模型需要通过一个简单的AVAL语言升级系统来学习任何新版本。