Existing Cross Modal Hashing (CMH) methods are mainly designed for balanced data, while imbalanced data with long-tail distribution is more general in real-world. Several long-tail hashing methods have been proposed but they can not adapt for multi-modal data, due to the complex interplay between labels and individuality and commonality information of multi-modal data. Furthermore, CMH methods mostly mine the commonality of multi-modal data to learn hash codes, which may override tail labels encoded by the individuality of respective modalities. In this paper, we propose LtCMH (Long-tail CMH) to handle imbalanced multi-modal data. LtCMH firstly adopts auto-encoders to mine the individuality and commonality of different modalities by minimizing the dependency between the individuality of respective modalities and by enhancing the commonality of these modalities. Then it dynamically combines the individuality and commonality with direct features extracted from respective modalities to create meta features that enrich the representation of tail labels, and binaries meta features to generate hash codes. LtCMH significantly outperforms state-of-the-art baselines on long-tail datasets and holds a better (or comparable) performance on datasets with balanced labels.
翻译:现有跨模模哈希(CMH)方法主要为平衡数据而设计,而长尾分布的不平衡数据则在现实世界中较为普遍。提出了几种长尾散列方法,但由于标签与个性和多模式数据共同信息之间的复杂相互作用,无法适应多模式数据。此外,CMH方法主要覆盖了多种模式数据的共同性,以学习散列码,而散列码可能取代由各自方式的个别性所编码的尾标签。在本文件中,我们提议LtCMH(Long-tail CMH)处理不平衡的多模式数据。LtCMH首先采用自动编码来清除不同模式的个别性和共性,尽量减少各自模式的个别性与共性之间的依赖性,并增强这些模式的共性。然后,将个人性和共性与从各自模式中提取的直接特征动态地结合起来,以创造元性特征,丰富尾标签的体现,并产生集成的元特征。LtCMH(Long-lex-levelas-s)在可比较性标签基准上明显超越了数据的弹性。