While attributes have been widely used for person re-identification (Re-ID) that matches the same person images across disjoint camera views, they are used either as extra features or for performing multi-task learning to assist the image-image person matching task. However, how to find a set of person images according to a given attribute description, which is very practical in many surveillance applications, remains a rarely investigated cross-modal matching problem in Person Re-ID. In this work, we present this challenge and employ adversarial learning to formulate the attribute-image cross-modal person Re-ID model. By imposing the regularization on the semantic consistency constraint across modalities, the adversarial learning enables generating image-analogous concepts for query attributes and getting it matched with image in both global level and semantic ID level. We conducted extensive experiments on three attribute datasets and demonstrated that the adversarial modelling is so far the most effective for the attributeimage cross-modal person Re-ID problem.
翻译:虽然属性被广泛用于个人再识别(Re-ID),这些属性在不连接的相机视图中与同一个人的图像相匹配,但它们要么用作额外特征,要么用于进行多任务学习,以协助图像图像图像人匹配任务。然而,如何根据一个特定属性描述找到一套个人图像,这在许多监视应用程序中非常实用,仍然是人再识别中很少调查的跨模式匹配问题。在这项工作中,我们提出这一挑战,并采用对抗性学习来制定属性图像模拟跨模式人再识别模式。通过对不同模式的语义一致性限制进行规范,对抗性学习能够生成用于查询属性的图像比喻概念,使其与全球级别和语义身份识别层面的图像相匹配。我们在三个属性数据集上进行了广泛的实验,并证明对抗性建模对于属性跨模式人再识别问题最为有效。