Text classification tends to be difficult when the data is deficient or when it is required to adapt to unseen classes. In such challenging scenarios, recent studies have often used meta-learning to simulate the few-shot task, thus negating explicit common linguistic features across tasks. Deep language representations have proven to be very effective forms of unsupervised pretraining, yielding contextualized features that capture linguistic properties and benefit downstream natural language understanding tasks. However, the effect of pretrained language representation for few-shot learning on text classification tasks is still not well understood. In this study, we design a few-shot learning model with pretrained language representations and report the empirical results. We show that our approach is not only simple but also produces state-of-the-art performance on a well-studied sentiment classification dataset. It can thus be further suggested that pretraining could be a promising solution for few shot learning of many other NLP tasks. The code and the dataset to replicate the experiments are made available at https://github.com/zxlzr/FewShotNLP.
翻译:当数据不足或需要适应不可见的类别时,文本分类往往很困难。在这种具有挑战性的情况下,最近的研究往往利用元学习模拟微小任务,从而淡化各种任务之间的明确共同语言特征。深层语言表现证明是未经监督的预培训非常有效的形式,产生反映语言特性并有利于下游自然语言理解任务的上下文特征。然而,对在文本分类任务上进行微粒学习的预先培训语言表现的影响仍然不甚了解。在这项研究中,我们设计了一个带有预先培训的语言表现的几发学习模型,并报告了经验性结果。我们表明,我们的方法不仅简单,而且还在经过仔细研究的情绪分类数据集上产生最先进的表现。因此,还可以进一步建议,预培训对于很少有机会了解许多其他NLP任务来说,可能是很有希望的解决办法。在https://github.com/zxlzr/FewShotNLP上提供复制实验的代码和数据集。