Scripts - standardized event sequences describing typical everyday activities - have been shown to help understand narratives by providing expectations, resolving ambiguity, and filling in unstated information. However, to date they have proved hard to author or extract from text. In this work, we demonstrate for the first time that pre-trained neural language models (LMs) can be be finetuned to generate high-quality scripts, at varying levels of granularity, for a wide range of everyday scenarios (e.g., bake a cake). To do this, we collected a large (6.4k), crowdsourced partially ordered scripts (named proScript), which is substantially larger than prior datasets, and developed models that generate scripts with combining language generation and structure prediction. We define two complementary tasks: (i) edge prediction: given a scenario and unordered events, organize the events into a valid (possibly partial-order) script, and (ii) script generation: given only a scenario, generate events and organize them into a (possibly partial-order) script. Our experiments show that our models perform well (e.g., F1=75.7 in task (i)), illustrating a new approach to overcoming previous barriers to script collection. We also show that there is still significant room for improvement toward human level performance. Together, our tasks, dataset, and models offer a new research direction for learning script knowledge.
翻译:演示了描述典型日常活动的标准化事件序列,通过提供期望、解决模棱两可和填充未注明的信息,帮助理解叙事。然而,迄今为止,它们证明对作者或从文本中摘录很难。在这项工作中,我们第一次展示了预先培训的神经语言模型(LMs)可以进行微调,以产生质量高的脚本,具有不同程度的颗粒性,用于广泛的日常情景(例如,烤蛋糕)。为了做到这一点,我们收集了一个大(6.4k)的、由众包制部分订购的脚本(名为ProScript),该脚本比先前的数据集大得多,并开发了生成脚本的模型,同时结合了语言生成和结构预测。我们定义了两个互补任务:(一) 边缘预测:根据一种假设和无序事件,将事件组织成一个有效的(可能是部分顺序的)脚本,以及(二) 脚本生成:仅给一种假想,生成事件,并将它们组织成一个(可能是部分顺序的)脚本。我们的实验显示我们的模型运行状况良好(e.g),F=收集了我们以前的脚本。