Meta-training, which fine-tunes the language model (LM) on various downstream tasks by maximizing the likelihood of the target label given the task instruction and input instance, has improved the zero-shot task generalization performance. However, meta-trained LMs still struggle to generalize to challenging tasks containing novel labels unseen during meta-training. In this paper, we propose Flipped Learning, an alternative method of meta-training which trains the LM to generate the task instruction given the input instance and label. During inference, the LM trained with Flipped Learning, referred to as Flipped, selects the label option that is most likely to generate the task instruction. On 14 tasks of the BIG-bench benchmark, the 11B-sized Flipped outperforms zero-shot T0-11B and even a 16 times larger 3-shot GPT-3 (175B) on average by 8.4% and 9.7% points, respectively. Flipped gives particularly large improvements on tasks with unseen labels, outperforming T0-11B by up to +20% average F1 score. This indicates that the strong task generalization of Flipped comes from improved generalization to novel labels. We release our code at https://github.com/seonghyeonye/Flipped-Learning.
翻译:在下游任务语言模型(LM)上微调语言模型(LM)的元培训,通过根据任务指令和输入实例最大限度地增加目标标签的可能性,微调语言模型(LM)在各种下游任务上的变化模式(LM),改进了零点任务一般化效果。然而,经过元培训的LM公司仍然在努力向包含在元培训过程中看不见的新标签的任务推广挑战性任务。在本文中,我们提议了Flipp Learning(Flippled Learning)的替代元培训方法,根据输入实例和标签来培训LM公司生成任务指示。在推断过程中,经过Flipp Learding培训的LM公司选择了最有可能产生任务指示的标签选项。在BIG-Bench基准的14项任务中, 11B规模的Fliflipped比零点T0-11B(甚至16倍以上GPT-PT-3(175B),分别平均8.4%和9.7%点。Flippled 特别改进了与TO-11B公司的任务一般化/Flishalshistationalationalationalation/We) 。这显示了我们GLIGIB公司/Fshishishistalshistaldaldation的强大任务一般化。