Black-box models, such as deep neural networks, exhibit superior predictive performances, but understanding their behavior is notoriously difficult. Many explainable artificial intelligence methods have been proposed to reveal the decision-making processes of black box models. However, their applications in high-stakes domains remain limited. Recently proposed neural additive models (NAM) have achieved state-of-the-art interpretable machine learning. NAM can provide straightforward interpretations with slight performance sacrifices compared with multi-layer perceptron. However, NAM can only model 1$^{\text{st}}$-order feature interactions; thus, it cannot capture the co-relationships between input features. To overcome this problem, we propose a novel interpretable machine learning method called higher-order neural additive models (HONAM) and a feature interaction method for high interpretability. HONAM can model arbitrary orders of feature interactions. Therefore, it can provide the high predictive performance and interpretability that high-stakes domains need. In addition, we propose a novel hidden unit to effectively learn sharp-shape functions. We conducted experiments using various real-world datasets to examine the effectiveness of HONAM. Furthermore, we demonstrate that HONAM can achieve fair AI with a slight performance sacrifice. The source code for HONAM is publicly available.
翻译:深神经网络等黑箱模型显示高超预测性能,但理解其行为却十分困难。许多可以解释的人工智能方法被提出来披露黑箱模型的决策过程。然而,在高吸收领域的应用仍然有限。最近提出的神经添加模型(NAM)已经实现了最先进的可解释机器学习。不结盟运动可以提供与多层透视系统相比的微小性能牺牲的简单解释。然而,不结盟运动只能提供1${text{st ⁇ $-order expective expact;因此,它无法捕捉输入功能之间的关联。为了克服这一问题,我们提出了一种新型的可解释的机器学习方法,称为更高级神经添加模型(HONAM)和高可解释性特征互动方法。HONAM可以建模特征互动的任意性命令。因此,它可以提供高水平的预测性能和可解释性能牺牲,而高层次透视系统需要。此外,我们提议建立一个新型的隐秘单位,以有效学习精锐的形状功能。我们用各种真实世界的数据集来进行实验,我们可以用各种现实世界的数据模型来检查国家空间数据库的效能。我们可以公开展示。