Recent studies show that large-scale sketch-based image retrieval (SBIR) can be efficiently tackled by cross-modal binary representation learning methods, where Hamming distance matching significantly speeds up the process of similarity search. Providing training and test data subjected to a fixed set of pre-defined categories, the cutting-edge SBIR and cross-modal hashing works obtain acceptable retrieval performance. However, most of the existing methods fail when the categories of query sketches have never been seen during training. In this paper, the above problem is briefed as a novel but realistic zero-shot SBIR hashing task. We elaborate the challenges of this special task and accordingly propose a zero-shot sketch-image hashing (ZSIH) model. An end-to-end three-network architecture is built, two of which are treated as the binary encoders. The third network mitigates the sketch-image heterogeneity and enhances the semantic relations among data by utilizing the Kronecker fusion layer and graph convolution, respectively. As an important part of ZSIH, we formulate a generative hashing scheme in reconstructing semantic knowledge representations for zero-shot retrieval. To the best of our knowledge, ZSIH is the first zero-shot hashing work suitable for SBIR and cross-modal search. Comprehensive experiments are conducted on two extended datasets, i.e., Sketchy and TU-Berlin with a novel zero-shot train-test split. The proposed model remarkably outperforms related works.
翻译:最近的研究显示,大规模草图图像检索(SBIR)可以通过跨模式的二进制图像学习方法(SBIR)来有效解决,在跨模式的二进制图像检索方法中,火腿距离相匹配大大加快了相似搜索过程。提供培训和测试数据时,要遵循一套固定的预设类别、尖端的SBIR和跨模式的散装散装材料才能取得可接受的检索性能。然而,在培训期间从未看到查询草图的类别时,大多数现有方法都失败了。在本文件中,上述问题被作为新颖的、但现实的零弹射模模模模的SBIR工作任务。我们详细介绍了这一特殊任务的挑战,并因此建议采用零发式的草图图像散装模型(ZSIH)模型(ZSIH)模型(ZSIH)模型(ZSIH)模型) 。一个端到端到端的三网络结构结构,其中两个是二进式的图象图像图像图像图像图像图像图像模型,这是我们提出的“零进化系统”的模型。我们设计了一个用于重建“S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-S-