We present SELOR, a framework for integrating self-explaining capabilities into a given deep model to achieve both high prediction performance and human precision. By "human precision", we refer to the degree to which humans agree with the reasons models provide for their predictions. Human precision affects user trust and allows users to collaborate closely with the model. We demonstrate that logic rule explanations naturally satisfy human precision with the expressive power required for good predictive performance. We then illustrate how to enable a deep model to predict and explain with logic rules. Our method does not require predefined logic rule sets or human annotations and can be learned efficiently and easily with widely-used deep learning modules in a differentiable way. Extensive experiments show that our method gives explanations closer to human decision logic than other methods while maintaining the performance of deep learning models.
翻译:我们提出SELOR,这是将自我解释能力纳入一个深层次模型以达到高预测性能和人类精确度的一个框架。我们通过“人类精确度”指的是人类同意模型提供预测的理由的程度。人类精确度影响用户信任度,并使用户能够与模型密切合作。我们证明逻辑规则解释自然地满足人类精确度,以良好的预测性能所需的表达力来满足人类精确度。然后我们说明如何使深层次模型能够预测和解释逻辑规则。我们的方法不需要预先界定的逻辑规则或人类说明,并且能够以不同的方式以广泛使用的深层学习模块来高效和容易地学习。广泛的实验表明,我们的方法比其他方法更接近人类决定性逻辑,同时保持深层次学习模型的性能。