Person re-identification (ReID) is a challenging crosscamera retrieval task to identify pedestrians. Many complex network structures are proposed recently and many of them concentrate on multi-branch features to achieve high performance. However, they are too heavy-weight to deploy in realworld applications. Additionally, pedestrian images are often captured by different surveillance cameras, so the varied lights, perspectives and resolutions result in inevitable multi-camera domain gaps for ReID. To address these issues, this paper proposes ATCN, a simple but effective angular triplet loss-based camera network, which is able to achieve compelling performance with only global features. In ATCN, a novel angular distance is introduced to learn a more discriminative feature representation in the embedding space. Meanwhile, a lightweight camera network is designed to transfer global features to more discriminative features. ATCN is designed to be simple and flexible so it can be easily deployed in practice. The experiment results on various benchmark datasets show that ATCN outperforms many SOTA approaches.
翻译:重新定位(ReID)是一项具有挑战性的跨相机检索任务,目的是识别行人。最近提出了许多复杂的网络结构,其中许多网络结构集中于多部门功能,以取得高性能。然而,它们过于重量,无法在现实世界应用中应用。此外,行人图像往往被不同的监视摄像机拍摄,因此不同的灯光、视角和分辨率为ReID带来了不可避免的多相机域空白。为了解决这些问题,本文件建议ATCN, 这是一个简单而有效的三角三重损失相机网络,能够实现仅具有全球特征的令人信服的性能。在ATCN, 引入了新型的角距离,以学习嵌入空间中更具歧视性的特征。与此同时,一个轻量相机网络旨在将全球特征转换为更具歧视性的特征。ATCN设计得既简单又灵活,便于实际应用。关于各种基准数据集的实验结果显示,ATCN优于许多SOTA方法。