Interpretability of neural networks and their underlying theoretical behaviour remain an open field of study even after the great success of their practical applications, particularly with the emergence of deep learning. In this work, NN2Poly is proposed: a theoretical approach to obtain an explicit polynomial model that provides an accurate representation of an already trained fully-connected feed-forward artificial neural network (a multilayer perceptron or MLP). This approach extends a previous idea proposed in the literature, which was limited to single hidden layer networks, to work with arbitrarily deep MLPs in both regression and classification tasks. The objective of this paper is to achieve this by using a Taylor expansion on the activation function, at each layer, and then using several combinatorial properties to calculate the coefficients of the desired polynomials. Discussion is presented on the main computational challenges of this method, and the way to overcome them by imposing certain constraints during the training phase. Finally, simulation experiments as well as an application to a real data set are presented to demonstrate the effectiveness of the proposed method.
翻译:神经网络及其基本理论行为即使在实际应用取得巨大成功之后,特别是随着深层学习的出现,其解释性仍然是开放的研究领域。在这项工作中,提出了NNN2Polly:一种理论方法,以获得一个明确的多元模型,准确反映一个已经受过训练的完全连接的进取-前进人工神经网络(多层感应器或MLP),这一方法扩展了文献中先前提出的一个概念,它仅限于单一的隐藏层网络,在回归和分类任务中与任意深度的MLP公司合作。本文件的目的是通过在每一层的激活功能上使用泰勒扩展,然后使用若干组合特性来计算预期的多层神经网络的系数,从而实现这一目标。讨论了这种方法的主要计算挑战,以及如何通过在培训阶段施加某些限制来克服这些挑战。最后,模拟实验和对真实数据集的应用展示了拟议方法的有效性。