Deep hashing has been widely applied to large-scale image retrieval tasks owing to efficient computation and low storage cost by encoding high-dimensional image data into binary codes. Since binary codes do not contain as much information as float features, the essence of binary encoding is preserving the main context to guarantee retrieval quality. However, the existing hashing methods have great limitations on suppressing redundant background information and accurately encoding from Euclidean space to Hamming space by a simple sign function. In order to solve these problems, a Cross-Scale Context Extracted Hashing Network (CSCE-Net) is proposed in this paper. Firstly, we design a two-branch framework to capture fine-grained local information while maintaining high-level global semantic information. Besides, Attention guided Information Extraction module (AIE) is introduced between two branches, which suppresses areas of low context information cooperated with global sliding windows. Unlike previous methods, our CSCE-Net learns a content-related Dynamic Sign Function (DSF) to replace the original simple sign function. Therefore, the proposed CSCE-Net is context-sensitive and able to perform well on accurate image binary encoding. We further demonstrate that our CSCE-Net is superior to the existing hashing methods, which improves retrieval performance on standard benchmarks.
翻译:由于将高维图像数据编码为二进制代码,将高维图像数据编码成二进制代码,从而降低存储成本,从而对大型图像检索任务广泛应用深层散列。由于二进制代码不包含与浮动特性同等多的信息,因此二进制编码的精髓保留了主要环境以保证检索质量。然而,现有的散列方法在抑制多余的背景资料和从欧clidean空间将低背景信息与全球滑动窗口合作的领域中存在着很大的局限性。与以往的方法不同,我们的CSCE-Net学习了与内容相关的动态信号函数(DSF)以取代原有的简单信号功能。因此,拟议的CSCE-Net设计了一个双权框架,以捕捉精细的本地信息,同时保持高层次的全球语义信息。此外,在两个分支之间引入了关注引导信息提取模块(AIE),以抑制与全球滑动窗口合作的低背景信息领域。与以往的方法不同,我们的CSCE-Net学习了与内容相关的动态信号函数(DSF),以取代原有的简单信号功能。因此,拟议的CSCE-Net框架对背景有敏感认识,并且能够对我们的准确的图像进行升级。