Explaining the black-box predictions of NLP models naturally and accurately is an important open problem in natural language generation. These free-text explanations are expected to contain sufficient and carefully-selected evidence to form supportive arguments for predictions. Due to the superior generative capacity of large pretrained language models, recent work built on prompt engineering enables explanation generation without specific training. However, explanation generated through single-pass prompting often lacks sufficiency and conciseness. To address this problem, we develop an information bottleneck method EIB to produce refined explanations that are sufficient and concise. Our approach regenerates the free-text explanation by polishing the single-pass output from the pretrained language model but retaining the information that supports the contents being explained. Experiments on two out-of-domain tasks verify the effectiveness of EIB through automatic evaluation and thoroughly-conducted human evaluation.
翻译:自然和准确地解释NLP模型的黑匣子预测自然和准确是自然语言生成中一个重要的未决问题。这些自由文本解释预计将包含足够和仔细选择的证据,以形成支持预测的论据。由于大型预先培训的语言模型的优异基因能力,最近基于迅速工程的工作使得无需具体培训就能够产生解释。然而,通过单方推动得出的解释往往缺乏充足性和简洁性。为了解决这一问题,我们开发了信息瓶颈法EIB, 以提出足够和简明的精细解释。我们的方法通过对预先培训的语言模型的单通文本输出进行擦亮,重新生成了自由文本解释,但保留了支持所解释内容的信息。关于两项外部任务的实验通过自动评估和彻底进行的人类评估来验证EIB的有效性。