A recent line of work showed that various forms of convolutional kernel methods can be competitive with standard supervised deep convolutional networks on datasets like CIFAR-10, obtaining accuracies in the range of 87-90% while being more amenable to theoretical analysis. In this work, we highlight the importance of a data-dependent feature extraction step that is key to the obtain good performance in convolutional kernel methods. This step typically corresponds to a whitened dictionary of patches, and gives rise to a data-driven convolutional kernel methods. We extensively study its effect, demonstrating it is the key ingredient for high performance of these methods. Specifically, we show that one of the simplest instances of such kernel methods, based on a single layer of image patches followed by a linear classifier is already obtaining classification accuracies on CIFAR-10 in the same range as previous more sophisticated convolutional kernel methods. We scale this method to the challenging ImageNet dataset, showing such a simple approach can exceed all existing non-learned representation methods. This is a new baseline for object recognition without representation learning methods, that initiates the investigation of convolutional kernel models on ImageNet. We conduct experiments to analyze the dictionary that we used, our ablations showing they exhibit low-dimensional properties.
翻译:最近一行工作表明,各种形式的革命内核方法可以与诸如CIFAR-10等标准、受监督的关于数据集的深层革命网络竞争,获得87-90%的精密度,同时更容易进行理论分析。在这项工作中,我们强调数据依赖特征提取步骤的重要性,这是在共生内核方法中取得良好性能的关键。这一步骤通常与白化的补丁字典相对应,并产生一种数据驱动的内核方法。我们广泛研究其效果,证明它是这些方法高性能的关键成分。具体地说,我们显示,基于单一一层图像补丁,这种内核方法的最简单的例子之一,已经获得CIRF-10的分类,与先前更复杂的革命内核方法相同。我们将这种方法推广到具有挑战性的图像网络数据集,显示这种简单的方法可以超越所有现有的非学习方法。这是在不进行展示性能识别对象方面,而没有展示这些方法的关键要素的关键要素。具体地说,我们显示,基于线性分类师所遵循的单一层图像补补补补补补,正在进行我们所进行低式的图像的模型。