ROUGE is a standard automatic evaluation metric based on n-grams for sequence-to-sequence tasks, while cross-entropy loss is an essential objective of neural network language model that optimizes at a unigram level. We present differentiable n-gram objectives, attempting to alleviate the discrepancy between training criterion and evaluating criterion. The objective maximizes the probabilistic weight of matched sub-sequences, and the novelty of our work is the objective weights the matched sub-sequences equally and does not ceil the number of matched sub-sequences by the ground truth count of n-grams in reference sequence. We jointly optimize cross-entropy loss and the proposed objective, providing decent ROUGE score enhancement over abstractive summarization dataset CNN/DM and XSum, outperforming alternative n-gram objectives.
翻译:ROUGE是一种标准的自动评价指标,基于序列至序列任务正克,而交叉作物损失是神经网络语言模型的一个基本目标,该模型在单克水平上优化。我们提出了不同的n克目标,试图缩小培训标准与评估标准之间的差异。目标是最大限度地提高匹配子序列数据的概率比重,我们工作的新颖之处是匹配子序列的客观加权相等,而不是在参考序列中以地面对正克真理计数为匹配子序列的数量。我们共同优化交叉作物损失和拟议目标,在CNN/DM和XSum的抽象合成数据集、超前的替代n克目标上提供体面的ROUGE分数增分。