Many methods now exist for conditioning model outputs on task instructions, retrieved documents, and user-provided explanations and feedback. Rather than relying solely on examples of task inputs and outputs, these approaches use valuable additional data for improving model correctness and aligning learned models with human priors. Meanwhile, a growing body of evidence suggests that some language models can (1) store a large amount of knowledge in their parameters, and (2) perform inference over tasks in textual inputs at test time. These results raise the possibility that, for some tasks, humans cannot explain to a model any more about the task than it already knows or could infer on its own. In this paper, we study the circumstances under which explanations of individual data points can (or cannot) improve modeling performance. In order to carefully control important properties of the data and explanations, we introduce a synthetic dataset for experiments, and we also make use of three existing datasets with explanations: e-SNLI, TACRED, and SemEval. We first give a formal framework for the available modeling approaches, in which explanation data can be used as model inputs, as targets, or as a prior. After arguing that the most promising role for explanation data is as model inputs, we propose to use a retrieval-based method and show that it solves our synthetic task with accuracies upwards of 95%, while baselines without explanation data achieve below 65% accuracy. We then identify properties of datasets for which retrieval-based modeling fails. With the three existing datasets, we find no improvements from explanation retrieval. Drawing on findings from our synthetic task, we suggest that at least one of six preconditions for successful modeling fails to hold with these datasets. Our code is publicly available at https://github.com/peterbhase/ExplanationRoles
翻译:现在有许多方法可以用来根据任务指示、检索到的文件和用户提供的解释和反馈调整模型产出。这些方法不仅依靠任务投入和产出的例子,而且使用宝贵的额外数据来改进模型的正确性,使学习到的模型与人类的前科相协调。与此同时,越来越多的证据表明,一些语言模型可以(1) 储存大量的参数知识,(2) 在测试时对文本输入中的任务进行推断。这些结果使人们有可能在一些任务中无法向模型解释任何比它自己已经知道或可以推断的关于任务的解释。在本文件中,我们研究解释单个数据点可以(或不能)改进模型性能的有价值的额外数据。为了仔细控制数据和解释的重要特性,我们引入了一个合成数据集用于实验,我们还使用三个现有的数据集来解释:e-SNLI、TACRED和SemEval。我们首先为现有的模型方法确定一个正式框架,在这个模型中,可以将数据解释作为模型,作为指标,或者作为最低指标,或者作为以前的模型,我们研究各个数据点的解释,在提出最有希望的数据的精确性的数据解释之后,我们用一个模型来解释。我们的数据解释,然后用一个可靠的数据模型来解释。我们用一个数据检索模型来显示一个数据解释,然后用一个具有的模型来解释。我们现有的数据的模型来显示一个令人承诺的精确性的数据解释。