Building explainable systems is a critical problem in the field of Natural Language Processing (NLP), since most machine learning models provide no explanations for the predictions. Existing approaches for explainable machine learning systems tend to focus on interpreting the outputs or the connections between inputs and outputs. However, the fine-grained information is often ignored, and the systems do not explicitly generate the human-readable explanations. To better alleviate this problem, we propose a novel generative explanation framework that learns to make classification decisions and generate fine-grained explanations at the same time. More specifically, we introduce the explainable factor and the minimum risk training approach that learn to generate more reasonable explanations. We construct two new datasets that contain summaries, rating scores, and fine-grained reasons. We conduct experiments on both datasets, comparing with several strong neural network baseline systems. Experimental results show that our method surpasses all baselines on both datasets, and is able to generate concise explanations at the same time.
翻译:建设可解释的系统是自然语言处理(NLP)领域的一个关键问题,因为大多数机器学习模式都没有对预测作出解释。现有可解释的机器学习系统的方法往往侧重于解释产出或投入与产出之间的联系。然而,微小的信息往往被忽视,而系统没有明确地产生人类可读的解释。为了更好地缓解这一问题,我们提出了一个创新的基因化解释框架,既学会作出分类决定,又同时作出精细的解释。更具体地说,我们引入了可解释的因素和最低风险培训方法,学会提出更合理的解释。我们建造了两个新的数据集,其中包含摘要、评级分数和细细细的分类理由。我们在两个数据集上进行实验,与几个强大的神经网络基线系统进行比较。实验结果表明,我们的方法超过了两个数据集的所有基线,同时能够产生简明的解释。