Person Re-identification (re-id) faces two major challenges: the lack of cross-view paired training data and learning discriminative identity-sensitive and view-invariant features in the presence of large pose variations. In this work, we address both problems by proposing a novel deep person image generation model for synthesizing realistic person images conditional on the pose. The model is based on a generative adversarial network (GAN) designed specifically for pose normalization in re-id, thus termed pose-normalization GAN (PN-GAN). With the synthesized images, we can learn a new type of deep re-id feature free of the influence of pose variations. We show that this feature is strong on its own and complementary to features learned with the original images. Importantly, under the transfer learning setting, we show that our model generalizes well to any new re-id dataset without the need for collecting any training data for model fine-tuning. The model thus has the potential to make re-id model truly scalable.
翻译:重新定位(重新定位)面临两大挑战:缺乏交叉视图对齐的培训数据,在存在巨大差异的情况下学习歧视性身份识别敏感和视异特征。在这项工作中,我们提出一个新的深层人图像生成模型,以综合以外观为条件的现实个人图像,以解决这两个问题。该模型基于一个基因对抗网络(GAN),专门为在重新定位中实现正常化而设计,因此被称为“面貌正常化GAN(PN-GAN) ” 。在合成图像中,我们可以学习一种新型的深层重新定位特征,而不受外观变异的影响。我们表明这一特征本身很强,与原始图像所学习的特征相辅相成。重要的是,在转移学习环境中,我们显示我们的模型非常适合任何新的重新定位数据集,无需为模型微调收集任何培训数据。因此,该模型有可能使重新定位模型真正具有可扩展性。