As the discipline has evolved, research in machine learning has been focused more and more on creating more powerful neural networks, without regard for the interpretability of these networks. Such "black-box models" yield state-of-the-art results, but we cannot understand why they make a particular decision or prediction. Sometimes this is acceptable, but often it is not. We propose a novel architecture, Regression Networks, which combines the power of neural networks with the understandability of regression analysis. While some methods for combining these exist in the literature, our architecture generalizes these approaches by taking interactions into account, offering the power of a dense neural network without forsaking interpretability. We demonstrate that the models exceed the state-of-the-art performance of interpretable models on several benchmark datasets, matching the power of a dense neural network. Finally, we discuss how these techniques can be generalized to other neural architectures, such as convolutional and recurrent neural networks.
翻译:随着学科的演进,机器学习的研究越来越侧重于创建更强大的神经网络,而没有考虑到这些网络的可解释性。这样的“黑盒模型”产生最先进的结果,但我们不明白为什么它们做出特定决定或预测。有时这是可以接受的,但往往并非如此。我们提出了一个新颖的结构,即回退网络,它将神经网络的力量与回归分析的可理解性结合起来。虽然文献中存在一些将这些结合的方法,但我们的建筑将这些方法概括化,方法是考虑到相互作用,提供稠密神经网络的力量,而不回避解释性。我们证明这些模型超过了几个基准数据集上可解释模型的最先进的性能,与稠密神经网络的力量相匹配。最后,我们讨论如何将这些技术推广到其他神经结构中,例如革命网络和经常性神经网络。