A fundamental bottleneck in utilising complex machine learning systems for critical applications has been not knowing why they do and what they do, thus preventing the development of any crucial safety protocols. To date, no method exist that can provide full insight into the granularity of the neural network's decision process. In the past, saliency maps were an early attempt at resolving this problem through sensitivity calculations, whereby dimensions of a data point are selected based on how sensitive the output of the system is to them. However, the success of saliency maps has been at best limited, mainly due to the fact that they interpret the underlying learning system through a linear approximation. We present a novel class of methods for generating nonlinear saliency maps which fully account for the nonlinearity of the underlying learning system. While agreeing with linear saliency maps on simple problems where linear saliency maps are correct, they clearly identify more specific drivers of classification on complex examples where nonlinearities are more pronounced. This new class of methods significantly aids interpretability of deep neural networks and related machine learning systems. Crucially, they provide a starting point for their more broad use in serious applications, where 'why' is equally important as 'what'.
翻译:在使用复杂机器学习系统进行关键应用方面,一个根本的瓶颈是不知道为什么使用复杂的机器学习系统,以及它们做了哪些工作,从而阻止了任何关键的安全协议的开发。迄今为止,还没有一种方法能够充分洞察神经网络决策过程的颗粒。过去,突出的地图是早期试图通过敏感性计算来解决这一问题,根据数据点的尺寸根据系统输出对它们的敏感性来选择数据点。然而,突出的地图的成败最多有限,主要是因为它们通过线性近似来解释基本学习系统。我们提出了一个新颖的方法,用来制作非线性突出的地图,充分说明基本学习系统的不直线性。在线性突出的地图正确的情况下,它们与线性突出的地图一样,在简单问题上,在非线性地图正确的情况下,它们清楚地确定了更具体的分类驱动因素,在非线性较明显的复杂例子中,这种新的方法非常有助于深海神经网络和相关机器学习系统的可解释性。毫无疑问,它们提供了一个起点,在严肃应用中,“为什么”同样重要。