Novel neural architectures, training strategies, and the availability of large-scale corpora haven been the driving force behind recent progress in abstractive text summarization. However, due to the black-box nature of neural models, uninformative evaluation metrics, and scarce tooling for model and data analysis, the true performance and failure modes of summarization models remain largely unknown. To address this limitation, we introduce SummVis, an open-source tool for visualizing abstractive summaries that enables fine-grained analysis of the models, data, and evaluation metrics associated with text summarization. Through its lexical and semantic visualizations, the tools offers an easy entry point for in-depth model prediction exploration across important dimensions such as factual consistency or abstractiveness. The tool together with several pre-computed model outputs is available at https://github.com/robustness-gym/summvis.
翻译:然而,由于神经模型的黑盒性质、无信息评价指标以及模型和数据分析的稀缺工具,对模型和数据分析的真正性能和失灵模式仍基本不为人所知。为了应对这一限制,我们引入了SummVis,这是一个可视化抽象摘要的开放源码工具,它有助于对与文本汇总有关的模型、数据和评价指标进行精细分析。工具通过它的词汇和语义可视化,为深入模型预测提供了一个容易的切入点,在诸如事实一致性或抽象性等重要方面进行深入的模型预测探索。工具以及若干预建模型产出可在https://github.com/robustness-gym/ sumvis查阅。