Generative modeling has been the dominant approach for large-scale pretraining and zero-shot generalization. In this work, we challenge this convention by showing that discriminative approaches perform substantially better than generative ones on a large number of NLP tasks. Technically, we train a single discriminator to predict whether a text sample comes from the true data distribution, similar to GANs. Since many NLP tasks can be formulated as selecting from a few options, we use this discriminator to predict the option with the highest probability. This simple formulation achieves state-of-the-art zero-shot results on the T0 benchmark, outperforming T0 by 16.0\%, 7.8\%, and 11.5\% respectively on different scales. In the finetuning setting, our approach also achieves new state-of-the-art results on a wide range of NLP tasks, with only 1/4 parameters of previous methods. Meanwhile, our approach requires minimal prompting efforts, which largely improves robustness and is essential for real-world applications. Furthermore, we also jointly train a generalized UD in combination with generative tasks, which maintains its advantage on discriminative tasks and simultaneously works on generative tasks.
翻译:在这项工作中,我们通过显示在大量NLP任务中,歧视性做法比基因化方法的效果要好得多。在技术上,我们培训了一个单一的导师,以预测文本样本是否来自真实的数据分布,类似于GANs。由于许多NLP任务可以从几个选项中选择,因此我们用这个导师来预测选项的概率最高。这一简单配方在T0基准上取得了最先进的零效果,在不同的尺度上分别超过T0+16.0、7.8+和11.5+。在微调设置中,我们的方法还在广泛的NLP任务中实现了新的最新结果,只有四分之一的先前方法参数。与此同时,我们的方法需要最低限度的迅速努力,这在很大程度上提高了稳健性,对于现实世界的应用至关重要。此外,我们还联合培训一个通用UD,同时进行基因化任务,在歧视任务上保持优势,同时进行基因化工作。