Pooling operations, which can be calculated at low cost and serve as a linear or nonlinear transfer function for data reduction, are found in almost every modern neural network. Countless modern approaches have already tackled replacing the common maximum value selection and mean value operations, not to mention providing a function that allows different functions to be selected through changing parameters. Additional neural networks are used to estimate the parameters of these pooling functions.Consequently, pooling layers may require supplementary parameters to increase the complexity of the whole model. In this work, we show that one perceptron can already be used effectively as a pooling operation without increasing the complexity of the model. This kind of pooling allows for the integration of multi-layer neural networks directly into a model as a pooling operation by restructuring the data and, as a result, learnin complex pooling operations. We compare our approach to tensor convolution with strides as a pooling operation and show that our approach is both effective and reduces complexity. The restructuring of the data in combination with multiple perceptrons allows for our approach to be used for upscaling, which can then be utilized for transposed convolutions in semantic segmentation.
翻译:在几乎所有现代神经网络中,几乎每一个现代神经网络都发现有无数现代方法已经解决了取代共同的最大价值选择和平均价值操作的问题,更不用说提供了一种功能,允许通过变化参数选择不同的功能。还使用了更多的神经网络来估计这些集合功能的参数。因此,集合层可能需要补充参数来增加整个模型的复杂程度。在这项工作中,我们表明,一个端点已经可以有效地用作集合操作,而不会增加模型的复杂性。这种集中可以将多层神经网络直接作为集合操作纳入模型,通过重组数据,从而学习复杂的集合操作。我们将我们的方法与电磁进进化作为集合操作加以比较,并表明我们的方法既有效,又降低了复杂性。数据与多端点相结合的重组使得我们能够用于升级,然后可以用于在断层分区中转换变换变变变变变变。