We introduce ParaBLEU, a paraphrase representation learning model and evaluation metric for text generation. Unlike previous approaches, ParaBLEU learns to understand paraphrasis using generative conditioning as a pretraining objective. ParaBLEU correlates more strongly with human judgements than existing metrics, obtaining new state-of-the-art results on the 2017 WMT Metrics Shared Task. We show that our model is robust to data scarcity, exceeding previous state-of-the-art performance using only $50\%$ of the available training data and surpassing BLEU, ROUGE and METEOR with only $40$ labelled examples. Finally, we demonstrate that ParaBLEU can be used to conditionally generate novel paraphrases from a single demonstration, which we use to confirm our hypothesis that it learns abstract, generalized paraphrase representations.
翻译:与以往的做法不同,Parableux学会以基因调节为培训前目标来理解参数。 与现有指标相比,Parableux与人类判断的关系更加密切,在2017年WMT计量共享任务中获得了新的最新成果。我们用我们所用的假设来证实它学会抽象、通用的参数表达方式。 我们用我们用来证实我们的假设,即它学会了抽象、通用的参数表达方式。