In recent years, large pretrained models have been used in dialogue systems to improve successful task completion rates. However, lack of reasoning capabilities of dialogue platforms make it difficult to provide relevant and fluent responses, unless the designers of a conversational experience spend a considerable amount of time implementing these capabilities in external rule based modules. In this work, we propose a novel method to fine-tune pretrained transformer models such as Roberta and T5. to reason over a set of facts in a given dialogue context. Our method includes a synthetic data generation mechanism which helps the model learn logical relations, such as comparison between list of numerical values, inverse relations (and negation), inclusion and exclusion for categorical attributes, and application of a combination of attributes over both numerical and categorical values, and spoken form for numerical values, without need for additional training dataset. We show that the transformer based model can perform logical reasoning to answer questions when the dialogue context contains all the required information, otherwise it is able to extract appropriate constraints to pass to downstream components (e.g. a knowledge base) when partial information is available. We observe that transformer based models such as UnifiedQA-T5 can be fine-tuned to perform logical reasoning (such as numerical and categorical attributes' comparison) over attributes that been seen in training time (e.g., accuracy of 90\%+ for comparison of smaller than $k_{\max}$=5 values over heldout test dataset).
翻译:近年来,在对话系统中使用了大量预先培训的模型,以提高任务完成率;然而,由于对话平台缺乏推理能力,因此很难提供相关和流畅的反应,除非对话经验的设计者花大量时间在外部规则模块中落实这些能力;在这项工作中,我们提出了一个新颖的方法,在特定对话背景下,根据一系列事实对Roberta和T5等预先培训的变压器模型进行微调,以便理解一系列事实;我们的方法包括一个合成数据生成机制,帮助模型学习逻辑关系,例如数字值清单之间的比较、反向关系(和否定)、对绝对属性的包容和排除,以及将数值和绝对值的属性和口述形式结合起来,而不需要额外的培训数据集。我们表明,基于变压器模型可以进行逻辑推理,在对话环境包含所有必要信息时回答问题,否则在可获得部分信息时,可以将适当的限制传递给下游部分(例如知识库),例如数字值、反向关系(和否定)、绝对属性的包容性和排除,以及数字形式的属性的组合组合,比对数据进行精确的比较。