We investigate the geometric properties of the functions learned by trained ConvNets in the preactivation space of their convolutional layers, by performing an empirical study of hyperplane arrangements induced by a convolutional layer. We introduce statistics over the weights of a trained network to study local arrangements and relate them to the training dynamics. We observe that trained ConvNets show a significant statistical bias towards regular hyperplane configurations. Furthermore, we find that layers showing biased configurations are critical to validation performance for the architectures considered, trained on CIFAR10, CIFAR100 and ImageNet.
翻译:我们研究了训练后 ConvNets 在其卷积层的激活空间中学习到的函数的几何属性,通过对卷积层引入的超平面排列进行实证研究。我们介绍了一些统计方法来研究局部排列,并将它们与训练动态联系起来。我们观察到,训练过的 ConvNets 显示出对于正则超平面配置的显著统计偏移。此外,我们发现,对于考虑到CIFAR10,CIFAR100和ImageNet 的架构来说,具有偏见配置的层对于验证性能至关重要。