Visual storytelling (VST) is the task of generating a story paragraph that describes a given image sequence. Most existing storytelling approaches have evaluated their models using traditional natural language generation metrics like BLEU or CIDEr. However, such metrics based on n-gram matching tend to have poor correlation with human evaluation scores and do not explicitly consider other criteria necessary for storytelling such as sentence structure or topic coherence. Moreover, a single score is not enough to assess a story as it does not inform us about what specific errors were made by the model. In this paper, we propose 3 evaluation metrics sets that analyses which aspects we would look for in a good story: 1) visual grounding, 2) coherence, and 3) non-redundancy. We measure the reliability of our metric sets by analysing its correlation with human judgement scores on a sample of machine stories obtained from 4 state-of-the-arts models trained on the Visual Storytelling Dataset (VIST). Our metric sets outperforms other metrics on human correlation, and could be served as a learning based evaluation metric set that is complementary to existing rule-based metrics.
翻译:视觉叙事(VST)是生成描述特定图像序列的故事段落的任务。 多数现有叙事方法都使用传统自然语言生成量度(如BLEU或CIDER)来评估其模型。 但是,基于n-g的比对往往与人类评价得分不相干,而且没有明确地考虑用于叙事的其他必要标准,如句号结构或专题一致性。 此外,单分不足以评估一个故事,因为它没有告诉我们模型中的具体错误。 在本文中,我们提议了3套评价指标,用以分析我们在一个好故事中要寻找的方面:1) 视觉地基、2) 一致性和3) 非冗余性。我们通过分析从在视觉叙事数据集(VIST)上培训的4个状态模型中获得的机器故事样本中与人类判断分数的关联度来衡量我们的成套度值的可靠性。 我们的衡量标准比其他人类相关性指标要优于其他指标,并且可以用作基于学习的评价指标集,以补充现有的基于规则的衡量尺度。