We study regularization in the context of small sample-size learning with over-parameterized neural networks. Specifically, we shift focus from architectural properties, such as norms on the network weights, to properties of the internal representations before a linear classifier. Specifically, we impose a topological constraint on samples drawn from the probability measure induced in that space. This provably leads to mass concentration effects around the representations of training instances, i.e., a property beneficial for generalization. By leveraging previous work to impose topological constraints in a neural network setting, we provide empirical evidence (across various vision benchmarks) to support our claim for better generalization.
翻译:具体地说,我们把重点从建筑特性(例如网络重量规范)转移到线性分类师面前的内部陈述的特性。具体地说,我们对从该空间引出的可能性测量中提取的样本施加了地形限制。这可以明显地导致围绕培训实例的表述产生大规模集中效应,即有利于概括化的财产。通过利用先前的工作在神经网络设置中施加地形限制,我们提供了经验证据(跨越各种愿景基准)来支持我们关于更好地概括化的主张。