Convolutional neural networks (CNNs) have achieved superior performance but still lack clarity about the nature and properties of feature extraction. In this paper, by analyzing the sensitivity of neural networks to frequencies and scales, we find that neural networks not only have low- and medium-frequency biases but also prefer different frequency bands for different classes, and the scale of objects influences the preferred frequency bands. These observations lead to the hypothesis that neural networks must learn the ability to extract features at various scales and frequencies. To corroborate this hypothesis, we propose a network architecture based on Gaussian derivatives, which extracts features by constructing scale space and employing partial derivatives as local feature extraction operators to separate high-frequency information. This manually designed method of extracting features from different scales allows our GSSDNets to achieve comparable accuracy with vanilla networks on various datasets.
翻译:进化神经网络(CNNs)取得了优异的性能,但仍不清楚地物提取的性质和性质。在本文中,通过分析神经网络对频率和比例的敏感性,我们发现神经网络不仅具有中低频偏差,而且偏爱不同等级的不同频带,物体的规模也影响偏好频带。这些观察导致一种假设,即神经网络必须学会在各种规模和频率上提取特征的能力。为了证实这一假设,我们提议建立一个基于高斯衍生物的网络结构,通过建造规模空间并使用部分衍生物作为局部地物提取操作者分离高频信息来提取特征。这种人工设计的从不同尺度上提取特征的方法使得我们的GSSDNet在各种数据集上与香草网络实现相似的准确性。</s>