Amongst others, the adoption of Rectified Linear Units (ReLUs) is regarded as one of the ingredients of the success of deep learning. ReLU activation has been shown to mitigate the vanishing gradient issue, to encourage sparsity in the learned parameters, and to allow for efficient backpropagation. In this paper, we recognize that the non-linear behavior of the ReLU function gives rise to a natural clustering when the pattern of active neurons is considered. This observation helps to deepen the learning mechanism of the network; in fact, we demonstrate that, within each cluster, the network can be fully represented as an affine map. The consequence is that we are able to recover an explanation, in the form of feature importance, for the predictions done by the network to the instances belonging to the cluster. Therefore, the methodology we propose is able to increase the level of interpretability of a fully connected feedforward ReLU neural network, downstream from the fitting phase of the model, without altering the structure of the network. A simulation study and the empirical application to the Titanic dataset, show the capability of the method to bridge the gap between the algorithm optimization and the human understandability of the black box deep ReLU networks.
翻译:除其他外,采用校正线性单元(ReLUs)被认为是深层学习成功的要素之一。 reLU的激活已被证明可以缓解渐渐消失的梯度问题,鼓励学习参数的宽度,并允许有效的反向调整。在本文件中,我们认识到,在考虑活跃神经元的模式时,RELU函数的非线性行为会产生自然集群。这种观察有助于加深网络的学习机制;事实上,我们证明,在每个集群内,网络可以完全作为近似图示来代表。其结果是,我们能够以特征重要性的形式找到一个解释,说明网络对属于该集群的情况所作的预测。因此,我们建议的方法能够提高完全相连的进料向上RELU神经网络的可解释性水平,从模型的适当阶段下游到下游,同时不改变网络的结构。对Tanticic数据集进行模拟研究和实验应用,从而显示我们能够以特征重要性的形式对网络对属于该集群的情况进行解释,从而能够理解该网络的深层次算算法和MAR值之间的距离。