Visible-infrared person re-identification (VI Re-ID) aims to match person images between the visible and infrared modalities. Existing VI Re-ID methods mainly focus on extracting homogeneous structural relationships from a single image, while ignoring the heterogeneous correlation between cross-modality images. The homogenous and heterogeneous structured relationships are crucial to learning effective identity representation and cross-modality matching. In this paper, we separately model the homogenous structural relationship by a modality-specific graph within individual modality and then mine the heterogeneous structural correlation in these two modality-specific graphs. First, the homogeneous structured graph (HOSG) mines one-vs.-rest relation between an arbitrary node (local feature) and all the rest nodes within a visible or infrared image to learn effective identity representation. Second, to find cross-modality identity-consistent correspondence, the heterogeneous graph alignment module (HGAM) further measures the relational edge strength by route search between two-modality local node features. Third, we propose the cross-modality cross-correlation (CMCC) loss to extract the modality invariance in heterogeneous global graph representation. CMCC computes the mutual information between modalities and expels semantic redundancy. Extensive experiments on SYSU-MM01 and RegDB datasets demonstrate that our method outperforms state-of-the-arts with a gain of 13.73\% and 9.45\% Rank1/mAP. The code is available at https://github.com/fegnyujian/Homogeneous-and-Heterogeneous-Relational-Graph.
翻译:可见红外人重新定位(VI Rev1/ID) 的目的是将可见和红外模式之间的个人图像匹配起来。 现有的VI 重新识别方法主要侧重于从单一图像中提取同质结构关系,而忽略跨现代图像之间的千差万别关联。 同一和多样化结构关系对于学习有效的身份代表与跨模式匹配至关重要。 在本文中, 我们分别用单个模式中特定模式的图表来模拟同质结构关系, 然后在这两个特定模式的图表中消除差异性结构关系。 首先, 单一结构式图表(HOSG) 地雷一至五。 在任意节点(当地特征)和在可见或红外图像中的所有其余节点之间重现关系,以学习有效的身份代表。 其次, 要找到交叉模式身份一致和跨模式的图表协调模块(HGAM) 进一步通过两种模式搜索方式的本地节点特征特征特征来测量关系边缘强度。 第三, 我们提议在一个任意结构化图表(CMCC) 中将跨模式(CMAC) 损失用于提取可获取的当前数据格式模式。