Statistical learning theory provides bounds on the necessary number of training samples needed to reach a prescribed accuracy in a learning problem formulated over a given target class. This accuracy is typically measured in terms of a generalization error, that is, an expected value of a given loss function. However, for several applications -- for example in a security-critical context or for problems in the computational sciences -- accuracy in this sense is not sufficient. In such cases, one would like to have guarantees for high accuracy on every input value, that is, with respect to the uniform norm. In this paper we precisely quantify the number of training samples needed for any conceivable training algorithm to guarantee a given uniform accuracy on any learning problem formulated over target classes containing (or consisting of) ReLU neural networks of a prescribed architecture. We prove that, under very general assumptions, the minimal number of training samples for this task scales exponentially both in the depth and the input dimension of the network architecture. As a corollary we conclude that the training of ReLU neural networks to high uniform accuracy is intractable. In a security-critical context this points to the fact that deep learning based systems are prone to being fooled by a possible adversary. We corroborate our theoretical findings by numerical results.
翻译:统计学习理论提供了一定数量的培训样本的界限,这些样本对于在特定目标类别上拟订的学习问题达到一定的准确性是必要的。这种准确性通常用一个一般错误来衡量,即某一损失函数的预期值。但是,对于若干应用,例如安全临界环境或计算科学方面的问题,这种意义上的准确性是不够的。在这种情况下,人们希望保证每个输入值的高度准确性,即统一规范。在本文中,我们精确地量化任何可想象的培训算法所需的培训样本数量,以保证在包含(或包括)一个指定结构的ReLU神经网络的目标类别上拟订的任何学习问题具有一致的准确性。我们证明,根据非常笼统的假设,这一任务尺度的培训样本在深度和网络结构的投入层面都极小。我们的结论是,对ReLU神经网络进行高度统一准确性的培训是棘手的。在安全临界背景下,我们指出,基于深层次学习的系统很容易被一个可能的敌人所欺骗。我们用数字来证实我们的理论结果。