Determining the plausibility of causal relations between clauses is a commonsense reasoning task that requires complex inference ability. The general approach to this task is to train a large pretrained language model on a specific dataset. However, the available training data for the task is often scarce, which leads to instability of model training or reliance on the shallow features of the dataset. This paper presents a number of techniques for making models more robust in the domain of causal reasoning. Firstly, we perform adversarial training by generating perturbed inputs through synonym substitution. Secondly, based on a linguistic theory of discourse connectives, we perform data augmentation using a discourse parser for detecting causally linked clauses in large text, and a generative language model for generating distractors. Both methods boost model performance on the Choice of Plausible Alternatives (COPA) dataset, as well as on a Balanced COPA dataset, which is a modified version of the original data that has been developed to avoid superficial cues, leading to a more challenging benchmark. We show a statistically significant improvement in performance and robustness on both datasets, even with only a small number of additionally generated data points.
翻译:确定条款之间因果关系的可信度是一项常识推理任务,需要复杂的推理能力。这项任务的一般方法是在具体数据集上培训一个大型的预先培训语言模型。然而,这项任务的现有培训数据往往很少,导致示范培训不稳定,或依赖数据集的浅质特征。本文件介绍了使模型在因果推理领域更加稳健的若干技术。首先,我们进行对抗性培训,通过同义替代生成过敏输入。第二,根据语言学理论,我们利用一个谈话连接学分析器来进行数据增强,我们使用一个谈话分析器来探测大文本中的因果联系条款,以及生成分散器的基因化语言模型。这两种方法都促进了选择可变异替代数据集的模式性能,以及平衡的COPA数据集的性能,这是为避免浅色提示而开发的原始数据的修改版,导致更具有挑战性的基准。我们展示了两个数据集的性能和稳健性在统计上显著的改进,即使是少量的额外数据。