Explainable machine learning models primarily justify predicted labels using either extractive rationales (i.e., subsets of input features) or free-text natural language explanations (NLEs) as abstractive justifications. While NLEs can be more comprehensive than extractive rationales, machine-generated NLEs have been shown to sometimes lack commonsense knowledge. Here, we show that commonsense knowledge can act as a bridge between extractive rationales and NLEs, rendering both types of explanations better. More precisely, we introduce a unified framework, called RExC (Rationale-Inspired Explanations with Commonsense), that (1) extracts rationales as a set of features responsible for machine predictions, (2) expands the extractive rationales using available commonsense resources, and (3) uses the expanded knowledge to generate natural language explanations. Our framework surpasses by a large margin the previous state-of-the-art in generating NLEs across five tasks in both natural language processing and vision-language understanding, with human annotators consistently rating the explanations generated by RExC to be more comprehensive, grounded in commonsense, and overall preferred compared to previous state-of-the-art models. Moreover, our work shows that commonsense-grounded explanations can enhance both task performance and rationales extraction capabilities.
翻译:可解释的机器学习模型主要证明使用采掘理由(即投入特征子集)或自由文本自然语言解释(NLEs)作为抽象理由的预测标签是合理的。虽然国家环境指标可以比采掘理由更加全面,但机器产生的NLEs有时证明缺乏常识知识。在这里,我们表明,常识知识可以作为采掘理由和国家环境指标之间的桥梁,使这两种解释都更好。更确切地说,我们引入了一个统一框架,称为RExC(Recation-Inspirates with Compsense),即(1) 提取理由,作为一套负责机器预测的特征,(2) 利用现有的共同环境资源扩大采掘理由,(3) 利用扩大的知识来产生自然语言解释。 我们的框架大大超过先前在自然语言处理和愿景语言理解方面产生NLE的五个任务时的状态。 更精确地说,人类警告者一贯将RExC(RExC)产生的解释评为更全面、基于共同意识和总体首选的模型,以比我们以往的采掘理由模型都显示共同的绩效。