Even though it has extensively been shown that retrieval specific training of deep neural networks is beneficial for nearest neighbor image search quality, most of these models are trained and tested in the domain of landmarks images. However, some applications use images from various other domains and therefore need a network with good generalization properties - a general-purpose CBIR model. To the best of our knowledge, no testing protocol has so far been introduced to benchmark models with respect to general image retrieval quality. After analyzing popular image retrieval test sets we decided to manually curate GPR1200, an easy to use and accessible but challenging benchmark dataset with a broad range of image categories. This benchmark is subsequently used to evaluate various pretrained models of different architectures on their generalization qualities. We show that large-scale pretraining significantly improves retrieval performance and present experiments on how to further increase these properties by appropriate fine-tuning. With these promising results, we hope to increase interest in the research topic of general-purpose CBIR.
翻译:尽管广泛表明对深神经网络的检索具体培训有利于近邻图像搜索质量,但大多数这些模型都是在里程碑图像领域培训和测试的,但有些应用使用来自其他不同领域的图像,因此需要一个具有良好通用特性的网络――通用 CBIR 模型。据我们所知,迄今为止没有采用测试协议来为一般图像检索质量的模型进行基准测试。在分析大众图像检索测试组之后,我们决定手动整理GPR1200,这是一个易于使用和容易获取但具有挑战性的基准数据集,具有广泛的图像类别。这一基准随后用于评估不同结构的预培训模型的通用性。我们表明,大规模预培训可大大改进检索性,并就如何通过适当的微调进一步增加这些特性提出实验。通过这些有希望的结果,我们希望增加人们对通用 CBIR 研究课题的兴趣。