Self-rationalization models that predict task labels and generate free-text elaborations for their predictions could enable more intuitive interaction with NLP systems. These models are, however, currently trained with a large amount of human-written free-text explanations for each task which hinders their broader usage. We propose to study a more realistic setting of self-rationalization using few training examples. We present FEB -- a standardized collection of four existing English-language datasets and associated metrics. We identify the right prompting approach by extensively exploring natural language prompts on FEB. Then, by using this prompt and scaling the model size, we demonstrate that making progress on few-shot self-rationalization is possible. We show there is still ample room for improvement in this task: the average plausibility of generated explanations assessed by human annotators is at most 51% (with GPT-3), while plausibility of human explanations is 76%. We hope that FEB and our proposed approach will spur the community to take on the few-shot self-rationalization challenge.
翻译:预测任务标签的自定义模型和为其预测生成自由文本说明的自定义模型可以使与NLP系统进行更直观的互动。 但是,这些模型目前经过培训,对阻碍其更广泛使用的每一项任务都有大量的人写自由文本解释。 我们提议利用几个培训实例研究更现实的自定义设置。 我们介绍了FEB -- -- 现有四个英语数据集和相关指标的标准化汇编。我们通过广泛探索FEB的自然语言提示,确定了正确的促进方法。然后,通过利用这一迅速和缩小模型的大小,我们证明在少见的自标化方面取得进展是可能的。我们表明,在这项任务中仍有很大的改进余地:由人类通知员所评估的自标的平均合理性最多为51%(与GPT-3),而人类解释的可辨性为76 %。 我们希望,FEB和我们提出的方法将激励社区应对几发自标自标挑战。