Self-supervised learning approach like contrastive learning is attached great attention in natural language processing. It uses pairs of training data augmentations to build a classification task for an encoder with well representation ability. However, the construction of learning pairs over contrastive learning is much harder in NLP tasks. Previous works generate word-level changes to form pairs, but small transforms may cause notable changes on the meaning of sentences as the discrete and sparse nature of natural language. In this paper, adversarial training is performed to generate challenging and harder learning adversarial examples over the embedding space of NLP as learning pairs. Using contrastive learning improves the generalization ability of adversarial training because contrastive loss can uniform the sample distribution. And at the same time, adversarial training also enhances the robustness of contrastive learning. Two novel frameworks, supervised contrastive adversarial learning (SCAL) and unsupervised SCAL (USCAL), are proposed, which yields learning pairs by utilizing the adversarial training for contrastive learning. The label-based loss of supervised tasks is exploited to generate adversarial examples while unsupervised tasks bring contrastive loss. To validate the effectiveness of the proposed framework, we employ it to Transformer-based models for natural language understanding, sentence semantic textual similarity and adversarial learning tasks. Experimental results on GLUE benchmark tasks show that our fine-tuned supervised method outperforms BERT$_{base}$ over 1.75\%. We also evaluate our unsupervised method on semantic textual similarity (STS) tasks, and our method gets 77.29\% with BERT$_{base}$. The robustness of our approach conducts state-of-the-art results under multiple adversarial datasets on NLI tasks.
翻译:在自然语言处理中,对比式学习等自监督学习方法在自然语言处理中受到极大关注。 它使用一对培训数据增强器来为具有良好代表性的编码器建立分类任务。 但是, 在 NLP 任务中, 构建学习配对比对比式学习要困难得多。 以前的工作会给成对制带来字级变化, 但是小变换可能会对判决的含义产生显著变化, 因为它是自然语言的离散和稀疏性质。 在本文中, 进行对抗性培训是为了产生比NLP的嵌入空间更具有挑战性、 更难学习的对立式实例。 使用对比式学习可以提高对立式培训的常规化能力, 因为对比性损失可以统一样本分布。 同时, 对抗性培训也提高了对比性学习的强性。 有两个新框架, 监督性对抗性对抗性学习( SCL) 和不超前的SCAL (USCAL), 通过使用对立式的对立式培训来产生对正值的对等性对等性学习。 基于标签的任务的丢失了非对等性示例示例示例示例示例实例,, 也用来生成了我们用于对等式的对等式的对等式的对等式的对等式的对等性任务 。