Visible-Infrared Person Re-Identification (VI-ReID) is a challenging retrieval task under complex modality changes. Existing methods usually focus on extracting discriminative visual features while ignoring the reliability and commonality of visual features between different modalities. In this paper, we propose a novel deep learning framework named Progressive Modality-shared Transformer (PMT) for effective VI-ReID. To reduce the negative effect of modality gaps, we first take the gray-scale images as an auxiliary modality and propose a progressive learning strategy. Then, we propose a Modality-Shared Enhancement Loss (MSEL) to guide the model to explore more reliable identity information from modality-shared features. Finally, to cope with the problem of large intra-class differences and small inter-class differences, we propose a Discriminative Center Loss (DCL) combined with the MSEL to further improve the discrimination of reliable features. Extensive experiments on SYSU-MM01 and RegDB datasets show that our proposed framework performs better than most state-of-the-art methods. For model reproduction, we release the source code at https://github.com/hulu88/PMT.
翻译:在复杂的模式变化下,可见-红外线重新识别(VI-REID)是一项具有挑战性的检索任务。现有方法通常侧重于提取歧视性视觉特征,而忽视不同模式之间视觉特征的可靠性和共性。在本文件中,我们提出一个新的深层次学习框架,名为进步模式共享变异器(PMT),用于有效的VI-REID。为了减少模式差距的负面影响,我们首先将灰度图像作为一种辅助模式,并提议一个渐进学习战略。然后,我们提议一种模式-共享增强损失(MSEL),以指导从模式共享特征中探索更可靠的身份信息的模式。最后,为了应对大类内差异和小类间差异的问题,我们提议了一个差异中心差异(DCL),与MSEL相结合,以进一步改善对可靠特征的歧视。关于SYSU-MM01和RegDB数据集的广泛实验表明,我们提议的框架比大多数状态-艺术方法都表现得更好。关于复制,我们在 https://gimbh.com/com.