Although the self-supervised pre-training of transformer models has resulted in the revolutionizing of natural language processing (NLP) applications and the achievement of state-of-the-art results with regard to various benchmarks, this process is still vulnerable to small and imperceptible permutations originating from legitimate inputs. Intuitively, the representations should be similar in the feature space with subtle input permutations, while large variations occur with different meanings. This motivates us to investigate the learning of robust textual representation in a contrastive manner. However, it is non-trivial to obtain opposing semantic instances for textual samples. In this study, we propose a disentangled contrastive learning method that separately optimizes the uniformity and alignment of representations without negative sampling. Specifically, we introduce the concept of momentum representation consistency to align features and leverage power normalization while conforming the uniformity. Our experimental results for the NLP benchmarks demonstrate that our approach can obtain better results compared with the baselines, as well as achieve promising improvements with invariance tests and adversarial attacks. The code is available in https://github.com/zxlzr/DCL.
翻译:虽然由自我监督的变压器模型培训前的自我监督使自然语言处理(NLP)应用发生了革命性的变化,并取得了各种基准的最新结果,但这一进程仍然易受来自合法投入的微小和难以察觉的变异的影响,从直觉上看,在特征空间的表达方式应该相似,输入变化微妙,而差异很大,其含义不同。这促使我们以对比的方式调查对稳健的文本表达方式的学习情况。然而,为文本样本取得相反的语义学实例是非三重性的。在本研究中,我们提出了一个分解的对比学习方法,在不进行负面抽样的情况下,分别优化表述的一致性和一致性。具体地说,我们引入了动力代表一致性的概念,以在符合统一性的同时调整特征和利用权力正常化。我们关于NLP基准的实验结果表明,我们的方法与基线相比,可以取得更好的结果,并在差异测试和对抗性攻击方面实现有希望的改进。该代码可在 https://githbub.com/zrzrz/Lz/s中查阅。