In this paper, we explore how to utilize pre-trained language model to perform few-shot text classification where only a few annotated examples are given for each class. Since using traditional cross-entropy loss to fine-tune language model under this scenario causes serious overfitting and leads to sub-optimal generalization of model, we adopt supervised contrastive learning on few labeled data and consistency-regularization on vast unlabeled data. Moreover, we propose a novel contrastive consistency to further boost model performance and refine sentence representation. After conducting extensive experiments on four datasets, we demonstrate that our model (FTCC) can outperform state-of-the-art methods and has better robustness.
翻译:在本文中,我们探索了如何利用预先培训的语言模型来进行微小的文本分类,而每一类只提供几个附加说明的例子。由于使用传统的跨热带损失来微调这种假设情景下的语言模型导致严重超标,并导致模型的简单化,我们采取了监督的对比性学习方法,以了解少数有标签的数据和大量无标签数据的一致性。此外,我们提出了一种新的对比性一致性,以进一步提升模型的性能和完善句子表达方式。在对四个数据集进行广泛实验之后,我们证明我们的模型(CTCC)能够超越最先进的方法,并且更加可靠。