In this paper, we present a large scale unlabeled person re-identification (Re-ID) dataset "LUPerson" and make the first attempt of performing unsupervised pre-training for improving the generalization ability of the learned person Re-ID feature representation. This is to address the problem that all existing person Re-ID datasets are all of limited scale due to the costly effort required for data annotation. Previous research tries to leverage models pre-trained on ImageNet to mitigate the shortage of person Re-ID data but suffers from the large domain gap between ImageNet and person Re-ID data. LUPerson is an unlabeled dataset of 4M images of over 200K identities, which is 30X larger than the largest existing Re-ID dataset. It also covers a much diverse range of capturing environments (eg, camera settings, scenes, etc.). Based on this dataset, we systematically study the key factors for learning Re-ID features from two perspectives: data augmentation and contrastive loss. Unsupervised pre-training performed on this large-scale dataset effectively leads to a generic Re-ID feature that can benefit all existing person Re-ID methods. Using our pre-trained model in some basic frameworks, our methods achieve state-of-the-art results without bells and whistles on four widely used Re-ID datasets: CUHK03, Market1501, DukeMTMC, and MSMT17. Our results also show that the performance improvement is more significant on small-scale target datasets or under few-shot setting.
翻译:在本文中,我们展示了大规模无标签的人重新识别(Re-ID)数据集“LUPerson”,并首次尝试进行未经监督的预培训,以提高学习者重新识别特征代表的普及能力。这是为了解决所有现有的人重新识别数据集都因数据注释需要付出昂贵的努力而规模有限的问题。先前的研究试图利用在图像网上预先培训过的模型,以缓解人重新识别数据短缺,但因图像网与人重新识别数据之间的巨大域间差距而受到影响。LUPerson是一个未经监管的预培训数据集,由超过200K身份的4M图像组成,比现有最大的重新识别特征代表能力大30X倍。这还涉及各种各样的捕捉环境(例如,相机设置、场景等)。基于这一数据集,我们系统地研究从两个角度学习重新识别特征的关键因素:数据增强和对比损失。在这个大规模数据存储上进行的未经超前培训有效地导致一个通用的重新识别数据图像数据集,在常规重新标识的4个基准框架上,也可以在使用所有现有数据市场(例如,相机设置、图像设置)之前实现所有现有数据更新的数据结果。