Since the meaning representations are detailed and accurate annotations which express fine-grained sequence-level semtantics, it is usually hard to train discriminative semantic parsers via Maximum Likelihood Estimation (MLE) in an autoregressive fashion. In this paper, we propose a semantic-aware contrastive learning algorithm, which can learn to distinguish fine-grained meaning representations and take the overall sequence-level semantic into consideration. Specifically, a multi-level online sampling algorithm is proposed to sample confusing and diverse instances. Three semantic-aware similarity functions are designed to accurately measure the distance between meaning representations as a whole. And a ranked contrastive loss is proposed to pull the representations of the semantic-identical instances together and push negative instances away. Experiments on two standard datasets show that our approach achieves significant improvements over MLE baselines and gets state-of-the-art performances by simply applying semantic-aware contrastive learning on a vanilla Seq2Seq model.
翻译:由于含义表达方式是详细和准确的说明,表达细微的序列级测序模拟学,因此通常很难以自动递减的方式,通过最大相似度估测法来培训歧视性的语义分析师。在本文中,我们建议采用一种具有语义特征的对比性学习算法,这种算法可以学会区分细微分度表示法,并考虑整个序列级测义。具体地说,建议采用多层次的在线取样算法来抽样混淆和不同的实例。三种具有语义特征的类似功能旨在准确地测量整个含义表达方式之间的距离。还提出等级对比性损失,以将语义相似的事例的表达方式结合在一起,推开负面的事例。对两个标准数据集的实验表明,我们的方法通过在香草Seq2Sqeq模型上仅仅应用具有语义意识的对比性学习,就可以大大改进MLE的基线,并获得最先进的表现。