Script Knowledge (Schank and Abelson, 1975) has long been recognized as crucial for language understanding as it can help in filling in unstated information in a narrative. However, such knowledge is expensive to produce manually and difficult to induce from text due to reporting bias (Gordon and Van Durme, 2013). In this work, we are interested in the scientific question of whether explicit script knowledge is present and accessible through pre-trained generative language models (LMs). To this end, we introduce the task of generating full event sequence descriptions (ESDs) given a scenario in the form of natural language prompts. In zero-shot probing experiments, we find that generative LMs produce poor ESDs with mostly omitted, irrelevant, repeated or misordered events. To address this, we propose a pipeline-based script induction framework (SIF) which can generate good quality ESDs for unseen scenarios (e.g., bake a cake). SIF is a two-staged framework that fine-tunes LM on a small set of ESD examples in the first stage. In the second stage, ESD generated for an unseen scenario is post-processed using RoBERTa-based models to filter irrelevant events, remove repetitions, and reorder the temporally misordered events. Through automatic and manual evaluations, we demonstrate that SIF yields substantial improvements ($1$-$3$ BLUE points) over a fine-tuned LM. However, manual analysis shows that there is great room for improvement, offering a new research direction for inducing script knowledge.
翻译:文稿知识(Shank and Abelson, 1975年)长期以来被公认为对语言理解至关重要,因为它有助于在叙述中填充未说明的信息。然而,由于报告偏差(Gordon and Van Durme, 2013年),这种知识是人工制作的,难以从文本中诱发的,由于报告偏差(Gordon and Van Durme, 2013年),这种知识是昂贵的。在这项工作中,我们感兴趣的是一个科学问题,即是否存在明确的脚本知识,并通过经过事先训练的重塑语言模型(LMs)获得高质量的文稿知识。为此,我们引入了一个以自然语言提示为形式的情景生成完整事件序列描述(ESDs)的任务。在零发的演示实验中,我们发现基因化的LMs产生不好的ESDSDS, 在大部分省略、不相干、重复、重复或顺序错误的事件中,ESDSAAAAAA级的AVRAAAAAAAAA级的AA级分析过程是经过大量变现。