We study the generalization of deep learning models in relation to the convex hull of their training sets. A trained image classifier basically partitions its domain via decision boundaries and assigns a class to each of those partitions. The location of decision boundaries inside the convex hull of training set can be investigated in relation to the training samples. However, our analysis shows that in standard image classification datasets, all testing images are considerably outside that convex hull, in the pixel space, in the wavelet space, and in the internal representations learned by deep networks. Therefore, the performance of a trained model partially depends on how its decision boundaries are extended outside the convex hull of its training data. From this perspective which is not studied before, over-parameterization of deep learning models may be considered a necessity for shaping the extension of decision boundaries. At the same time, over-parameterization should be accompanied by a specific training regime, in order to yield a model that not only fits the training set, but also its decision boundaries extend desirably outside the convex hull. To illustrate this, we investigate the decision boundaries of a neural network, with various degrees of parameters, inside and outside the convex hull of its training set. Moreover, we use a polynomial decision boundary to study the necessity of over-parameterization and the influence of training regime in shaping its extensions outside the convex hull of training set.
翻译:我们研究深层学习模型的普及情况,这些模型与培训机群的骨架有关。受过训练的图像分类师基本上通过决定界限分割其域域,并给每个分区分配一个课级。训练机群内决定界限的位置可以与训练样品一起调查。然而,我们的分析表明,在标准图像分类数据集中,所有测试图像都相当地超出锥壳之外,在象素空间、波盘空间和深网络所学的内部演示中。因此,经过训练的模型的性能部分取决于其决定界限如何通过决定界限扩展到其训练数据的锥体外壳之外。从这个以前没有研究过的角度来看,深层学习模型的过分度可被视为形成决定界限延伸的必要性。与此同时,在标准图像分类数据集中,所有过量的图像都应伴有特定的训练制度,以便产生一个不仅适合训练设置的模型,而且其决定界限也明显延伸到锥体船体外的外壳体。为了说明这一点,我们从以前没有研究过研究过层结构网络的决定界限的界限,我们从外部研究了内测测测测测的界限范围,用了各种界限的界限的界限研究范围,我们内部测测测定的界限的界限内的界限的界限的界限的界限的界限的界限的界限范围,并定了我们内部的界限内定的界限的界限的界限的界限的界限范围。