In recent years, a growing body of research has focused on the problem of person re-identification (re-id). The re-id techniques attempt to match the images of pedestrians from disjoint non-overlapping camera views. A major challenge of re-id is the serious intra-class variations caused by changing viewpoints. To overcome this challenge, we propose a deep neural network-based framework which utilizes the view information in the feature extraction stage. The proposed framework learns a view-specific network for each camera view with a cross-view Euclidean constraint (CV-EC) and a cross-view center loss (CV-CL). We utilize CV-EC to decrease the margin of the features between diverse views and extend the center loss metric to a view-specific version to better adapt the re-id problem. Moreover, we propose an iterative algorithm to optimize the parameters of the view-specific networks from coarse to fine. The experiments demonstrate that our approach significantly improves the performance of the existing deep networks and outperforms the state-of-the-art methods on the VIPeR, CUHK01, CUHK03, SYSU-mReId, and Market-1501 benchmarks.
翻译:近些年来,越来越多的研究集中于重新确定人的身份(re-id)问题。重新确定的技术试图将行人图像与互不相连的非重叠相机视图相匹配。重新确定的一个主要挑战是由于观点的变化而导致的严重的阶级内部差异。为了克服这一挑战,我们提议了一个深神经网络框架,利用特征提取阶段的视觉信息。拟议框架学习了每个摄像头视景的视景网络,带有交叉视图的Euclidean限制(CV-EC)和交叉视图中心损失(CV-CL)。我们利用CV-EC来缩小不同观点之间特征的边际,并将中心损失指标扩展至特定视图版本,以更好地调整重新确定的问题。此外,我们提议了一个迭代算法,以优化特定视图网络的参数,从粗俗到精细。实验表明,我们的方法大大改进了现有深视网络的性能,超越了VENR、CUHK01、CUHK03、CUHK03、SIS-MIM-基准和SIS-DIS-SUM。