We propose a novel approach to summarization based on Bayesian deep learning. We approximate Bayesian summary generation by first extending state-of-the-art summarization models with Monte Carlo dropout and then using them to perform multiple stochastic forward passes. This method allows us to improve summarization performance by simply using the median of multiple stochastic summaries. We show that our variational equivalents of BART and PEGASUS can outperform their deterministic counterparts on multiple benchmark datasets. In addition, we rely on Bayesian inference to measure the uncertainty of the model when generating summaries. Having a reliable uncertainty measure, we can improve the experience of the end user by filtering out generated summaries of high uncertainty. Furthermore, our proposed metric could be used as a criterion for selecting samples for annotation, and can be paired nicely with active learning and human-in-the-loop approaches.
翻译:我们建议一种基于贝叶斯人深造的总结方法。 我们通过首次推广最先进的总结模型,与蒙特卡洛辍学者相近,然后利用这些模型进行多个随机前传。 这种方法使我们能够通过简单地使用多个随机总结的中位数来提高汇总性能。 我们显示,我们的BART和PEGASUS的变式等值在多个基准数据集上可以优于其确定性对应方。 此外,我们依靠贝叶斯人推论来测量模型在生成摘要时的不确定性。 有了可靠的不确定性测量,我们可以通过过滤生成的高度不确定性摘要来改善终端用户的经验。 此外,我们提议的衡量标准可以用作选择标注样本的标准,并且可以与积极学习和人行中方法相匹配。