Pre-trained Language Models (PTLMs) have been shown to perform well on natural language tasks. Many prior works have leveraged structured commonsense present in the form of entities linked through labeled relations in Knowledge Graphs (KGs) to assist PTLMs. Retrieval approaches use KG as a separate static module which limits coverage since KGs contain finite knowledge. Generative methods train PTLMs on KG triples to improve the scale at which knowledge can be obtained. However, training on symbolic KG entities limits their applicability in tasks involving natural language text where they ignore overall context. To mitigate this, we propose a CommonSense Contextualizer (CoSe-Co) conditioned on sentences as input to make it generically usable in tasks for generating knowledge relevant to the overall context of input text. To train CoSe-Co, we propose a novel dataset comprising of sentence and commonsense knowledge pairs. The knowledge inferred by CoSe-Co is diverse and contain novel entities not present in the underlying KG. We augment generated knowledge in Multi-Choice QA and Open-ended CommonSense Reasoning tasks leading to improvements over current best methods on CSQA, ARC, QASC and OBQA datasets. We also demonstrate its applicability in improving performance of a baseline model for paraphrase generation task.
翻译:预先培训的语言模型(PTLMs)在自然语言任务方面表现良好,许多以前的工作都利用了通过知识图(KGs)中标签关系联系的实体形式的结构化常识,以协助PTLMs。检索方法将KG作为一个单独的静态模块,限制覆盖面,因为KGs包含有限的知识。创新方法对PTLMs进行KG三重培训,以提高获得知识的规模。然而,关于象征性KG实体的培训限制了其在涉及自然语言文本的任务中的可适用性,而它们忽视了整体背景。为了减轻这一影响,我们提议以句子为条件,共同意识背景(COS-Co)为条件,共同思想背景(COS-C-Co)为生成与投入文本总体背景相关的知识的任务提供通用的可用性知识。为了培训COSE-C,我们提出了一套由句子和常识知识组合组成的新数据集。CSE-Coc 所推断的知识是多种多样的,并包含不在基本KG中的新实体。我们增加了在多方言组(QA)和开放共同语言应用应用性(CSQAQSQSQ)中产生知识的知识,我们还展示了在改进当前生成数据基准任务中的改进了目前生成数据的方法。