Image search can be tackled using deep features from pre-trained Convolutional Neural Networks (CNN). The feature map from the last convolutional layer of a CNN encodes descriptive information from which a discriminative global descriptor can be obtained. We propose a new representation of co-occurrences from deep convolutional features to extract additional relevant information from this last convolutional layer. Combining this co-occurrence map with the feature map, we achieve an improved image representation. We present two different methods to get the co-occurrence representation, the first one based on direct aggregation of activations, and the second one, based on a trainable co-occurrence representation. The image descriptors derived from our methodology improve the performance in very well-known image retrieval datasets as we prove in the experiments.
翻译:图像搜索可以使用预先培训的进化神经网络(CNN)的深层特征来解决。CNN编码描述性信息的最后进化层的地貌图,可以从中获取具有歧视性的全球描述性信息。我们建议对深进化特征的共同发生进行新的描述,以便从最后的进化层获取更多相关信息。将这一相联地图与地貌图结合起来,我们就能改善图像的描述。我们提出了两种不同的方法来获得共发式代表,一种是直接集聚激活数据,另一种是经过培训的共发性代表。我们的方法所产生的图像描述性能改进了我们在实验中证明的众所周知的图像检索数据集的性能。